Machine Learning Engineer Nanodegree

Reinforcement Learning

Project: Train a Smartcab to Drive

Welcome to the fourth project of the Machine Learning Engineer Nanodegree! In this notebook, template code has already been provided for you to aid in your analysis of the Smartcab and your implemented learning algorithm. You will not need to modify the included code beyond what is requested. There will be questions that you must answer which relate to the project and the visualizations provided in the notebook. Each section where you will answer a question is preceded by a 'Question X' header. Carefully read each question and provide thorough answers in the following text boxes that begin with 'Answer:'. Your project submission will be evaluated based on your answers to each of the questions and the implementation you provide in agent.py.

Note: Code and Markdown cells can be executed using the Shift + Enter keyboard shortcut. In addition, Markdown cells can be edited by typically double-clicking the cell to enter edit mode.


Getting Started

In this project, you will work towards constructing an optimized Q-Learning driving agent that will navigate a Smartcab through its environment towards a goal. Since the Smartcab is expected to drive passengers from one location to another, the driving agent will be evaluated on two very important metrics: Safety and Reliability. A driving agent that gets the Smartcab to its destination while running red lights or narrowly avoiding accidents would be considered unsafe. Similarly, a driving agent that frequently fails to reach the destination in time would be considered unreliable. Maximizing the driving agent's safety and reliability would ensure that Smartcabs have a permanent place in the transportation industry.

Safety and Reliability are measured using a letter-grade system as follows:

Grade Safety Reliability
A+ Agent commits no traffic violations,
and always chooses the correct action.
Agent reaches the destination in time
for 100% of trips.
A Agent commits few minor traffic violations,
such as failing to move on a green light.
Agent reaches the destination on time
for at least 90% of trips.
B Agent commits frequent minor traffic violations,
such as failing to move on a green light.
Agent reaches the destination on time
for at least 80% of trips.
C Agent commits at least one major traffic violation,
such as driving through a red light.
Agent reaches the destination on time
for at least 70% of trips.
D Agent causes at least one minor accident,
such as turning left on green with oncoming traffic.
Agent reaches the destination on time
for at least 60% of trips.
F Agent causes at least one major accident,
such as driving through a red light with cross-traffic.
Agent fails to reach the destination on time
for at least 60% of trips.

To assist evaluating these important metrics, you will need to load visualization code that will be used later on in the project. Run the code cell below to import this code which is required for your analysis.

In [1]:
# Import the visualization code
import visuals as vs

# Pretty display for notebooks
%matplotlib inline

Understand the World

Before starting to work on implementing your driving agent, it's necessary to first understand the world (environment) which the Smartcab and driving agent work in. One of the major components to building a self-learning agent is understanding the characteristics about the agent, which includes how the agent operates. To begin, simply run the agent.py agent code exactly how it is -- no need to make any additions whatsoever. Let the resulting simulation run for some time to see the various working components. Note that in the visual simulation (if enabled), the white vehicle is the Smartcab.

Question 1

In a few sentences, describe what you observe during the simulation when running the default agent.py agent code. Some things you could consider:

  • Does the Smartcab move at all during the simulation?
  • What kind of rewards is the driving agent receiving?
  • How does the light changing color affect the rewards?

Hint: From the /smartcab/ top-level directory (where this notebook is located), run the command

'python smartcab/agent.py'
  • Answer:
    1. The smartcab is not moving because the agent is not set to learn ie. Learning flag is set to False.
    2. The agent is not updated to take action so it remains idle irrespective of the traffic signal. Since it is idle in green signal where it supposed to perform action it is rewarded negative value and the value increases in negative side for violation of traffic rule.And it is rewarded positive value for remaining idle during red signal. When its voilate the rule ie making mistake it is penalized high by incresing reward value in negative side compare to increase of value in positive direction when it perform correct action.
In [10]:
%run smartcab/agent.py
/-------------------------
| Training trial 1
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.85)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 2.21)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.24)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 1.08)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.45)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.82)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.15)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.51)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.30)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.12)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 1.72)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.21)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.22)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 1.48)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.89)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.06)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.15)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.89)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.18)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.23)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.70)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.12)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.62)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded -0.52)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.20)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.35)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.27)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.63)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.15)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.74)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 2
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.64)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.99)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.27)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.27)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.47)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.83)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.19)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 1.33)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.30)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.34)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.09)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.51)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.66)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.56)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.22)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.59)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.42)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.95)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.08)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.86)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.90)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.66)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.71)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.10)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 3
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.12)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.41)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.43)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.47)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.31)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.73)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.18)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.32)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.27)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.40)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.32)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.58)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.44)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.47)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.42)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded -0.03)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.34)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded -0.23)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.40)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded -0.08)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.98)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.82)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 4
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.34)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.23)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.70)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.48)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.81)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.83)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 1.67)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.84)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.96)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.91)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.34)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.10)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.06)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.50)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 0.61)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.58)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded -0.50)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 5
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.98)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.85)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.65)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.85)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.15)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.22)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.31)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.21)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.94)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.06)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 0.91)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.48)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.37)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.08)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded -0.21)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.46)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.00)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.69)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.56)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.01)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.69)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.36)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.45)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded -0.15)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.64)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded -0.78)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 6
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.47)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.12)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.87)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.43)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.14)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.33)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.12)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.35)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.32)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.73)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 1.23)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.50)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.76)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.06)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.51)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.04)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.20)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.71)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.64)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.40)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 7
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 1.07)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.27)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.32)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.44)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.91)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.37)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.22)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.07)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.78)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.51)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.50)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.59)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.25)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 0.99)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 0.66)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.40)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.37)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.46)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.50)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 8
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.29)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 1.62)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.15)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.52)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.89)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.98)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.51)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.60)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.24)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.14)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.38)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.73)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.45)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.63)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.05)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.82)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.94)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.00)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.27)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.57)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.62)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.50)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.39)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 9
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.97)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.62)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.49)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.47)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.18)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.09)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.79)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 0.93)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.28)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 2.09)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.22)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.88)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.85)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.82)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.53)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.99)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded -0.64)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.64)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 10
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.34)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.94)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.27)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.03)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.13)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.23)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.36)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.19)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.62)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.41)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.56)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.00)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.34)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.86)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 0.54)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.50)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.01)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.12)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.58)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 11
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.67)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.66)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.81)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.70)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.48)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.33)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 1.36)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.93)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.22)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.28)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.86)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.87)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 2.37)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.32)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.10)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.15)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.79)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.60)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.00)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.03)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.99)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.92)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.43)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.89)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.12)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.82)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.66)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.26)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.80)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 12
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.00)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.14)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.84)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.32)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.45)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.67)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.04)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 2.92)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.70)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.29)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.46)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 2.37)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.69)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.19)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.80)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.19)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 2.20)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.37)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.02)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.37)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.43)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.24)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.42)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded -0.28)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.89)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.30)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.07)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.81)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 13
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.59)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 2.88)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.59)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.41)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.76)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.59)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.15)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.06)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.70)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.71)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.71)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.55)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.70)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.15)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.30)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.14)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.56)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.77)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.80)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 14
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.16)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.96)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.72)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.75)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.48)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.78)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.75)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.11)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.00)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.08)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.60)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.45)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.44)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.35)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.50)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.05)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.85)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.31)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.49)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.02)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.09)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.55)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.40)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.63)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 0.98)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 1.03)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.72)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded -0.56)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.36)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 15
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.90)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.02)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.79)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.63)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.01)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.65)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.55)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.10)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.30)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.53)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.49)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.60)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.05)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.92)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.68)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.42)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.68)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.87)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.68)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 16
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.36)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.93)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.95)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.09)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.59)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.09)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.52)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.12)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.66)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.93)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.45)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.86)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.96)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.05)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.82)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.67)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.30)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.14)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.57)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.93)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.18)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.24)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.20)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.27)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 17
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.09)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.53)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.55)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.27)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.79)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.05)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.49)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.95)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.04)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.51)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.51)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.10)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.12)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.93)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 2.12)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.36)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.88)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.43)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.03)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.19)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 0.48)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.63)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded -0.04)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 18
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.50)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.97)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.88)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.77)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.85)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.23)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.73)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.37)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 1.68)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.22)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.15)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 1.21)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.60)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.46)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.67)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.48)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.23)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.14)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.37)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.20)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.64)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.76)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.15)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.36)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 19
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.15)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.16)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.18)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.34)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.17)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.04)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.44)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.60)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.50)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.52)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.41)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.32)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.13)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.47)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded -0.43)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.28)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.01)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.44)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.17)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.63)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 20
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.00)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.64)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.99)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.39)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.12)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.09)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.83)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.54)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.55)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.62)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.80)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 2.55)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 2.66)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.11)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.31)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.37)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.33)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.04)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.94)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.46)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.72)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.04)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.39)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.81)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.17)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.30)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.60)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded -0.28)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.47)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 2.07)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.56)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.55)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.15)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.66)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 1
\-------------------------

Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.63)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.20)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.35)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.41)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.23)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.92)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.65)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.98)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.32)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.07)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.97)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.54)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.67)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 0.90)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded -0.37)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.86)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.19)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.20)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.76)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.14)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.75)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.95)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.54)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

Simulation ended. . . 

Answer:

Understand the Code

In addition to understanding the world, it is also necessary to understand the code itself that governs how the world, simulation, and so on operate. Attempting to create a driving agent would be difficult without having at least explored the "hidden" devices that make everything work. In the /smartcab/ top-level directory, there are two folders: /logs/ (which will be used later) and /smartcab/. Open the /smartcab/ folder and explore each Python file included, then answer the following question.

Question 2

  • In the agent.py Python file, choose three flags that can be set and explain how they change the simulation.
  • In the environment.py Python file, what Environment class function is called when an agent performs an action?
  • In the simulator.py Python file, what is the difference between the 'render_text()' function and the 'render()' function?
  • In the planner.py Python file, will the 'next_waypoint() function consider the North-South or East-West direction first?

Answer:

  • agent.py has three key flags:
  • learning (boolean): Whether the agent is expected to learn. This determines whether the agent will perform any Q-Learning (True), or instead simply choose random actions (False).
  • epsilon (float): Random exploration factor. This determines to what extent random actions will be performed. If this value is too low, the smartcab will not be able to learn new behaviour because it will be constantly drive safely and using previous learnings all the time. If this value is too high, then it will never be able to enact learnings and will constantly take risks. A value in between will balance previous learnings with risk-taking.
  • alpha (float): Learning factor. This defines the extent to which previous learnings are iterated on by new learnings. A value of 0 would mean that the algorithm would never change it's own behaviour. A value of 1 would be 'full learning', where no previous learnings are ever carried forward, and everything is 'forgotten' going forward. A value in between e.g. 0.5 would take an average of the previous value with the new learning.
  • environment.py: The act member function is called at the end of every update call within the agent. This requests the environment to provide a reward based upon the agent's action.
  • simulator.py has two rendering functions, render_text and render.The render_text() is incharge of providing the output of each step of the simulation to the command line / terminal .
  • while the render() method is incharge of drawing all the cars / lights / intersections and simulation information on to the pygame gui interface.window.
  • planner.py first seems to consider the East-West direction first by evaluating dx. It then evaluates the North-South direction by evaluating dy. There is no implementation for Anything in between - i.e. North-East, North-West etc.

Implement a Basic Driving Agent

The first step to creating an optimized Q-Learning driving agent is getting the agent to actually take valid actions. In this case, a valid action is one of None, (do nothing) 'left' (turn left), right' (turn right), or 'forward' (go forward). For your first implementation, navigate to the 'choose_action()' agent function and make the driving agent randomly choose one of these actions. Note that you have access to several class variables that will help you write this functionality, such as 'self.learning' and 'self.valid_actions'. Once implemented, run the agent file and simulation briefly to confirm that your driving agent is taking a random action each time step.

Basic Agent Simulation Results

To obtain results from the initial simulation, you will need to adjust following flags:

  • 'enforce_deadline' - Set this to True to force the driving agent to capture whether it reaches the destination in time.
  • 'update_delay' - Set this to a small value (such as 0.01) to reduce the time between steps in each trial.
  • 'log_metrics' - Set this to True to log the simluation results as a .csv file in /logs/.
  • 'n_test' - Set this to '10' to perform 10 testing trials.

Optionally, you may disable to the visual simulation (which can make the trials go faster) by setting the 'display' flag to False. Flags that have been set here should be returned to their default setting when debugging. It is important that you understand what each flag does and how it affects the simulation!

Once you have successfully completed the initial simulation (there should have been 20 training trials and 10 testing trials), run the code cell below to visualize the results. Note that log files are overwritten when identical simulations are run, so be careful with what log file is being loaded! Run the agent.py file after setting the flags from projects/smartcab folder instead of projects/smartcab/smartcab.

In [30]:
%run smartcab/agent.py
/-------------------------
| Training trial 1
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.55)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.24)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.46)
85% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.46)
85% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.46)
85% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.46)
85% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.46)
85% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.46)
85% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.46)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.44)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.42)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.40)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.08)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.51)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.74)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.74)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.74)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.74)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.74)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.74)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.74)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.74)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.74)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.74)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.39)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.15)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.56)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.35)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.34)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.86)
15% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.86)
15% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.86)
15% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.86)
15% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.86)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.91)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
5% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
5% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
5% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
5% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
5% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
5% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
5% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
5% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
5% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
5% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.79)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 2
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.85)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.96)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.59)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.75)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.75)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.75)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.75)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.75)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.34)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.83)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.82)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.38)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.67)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.67)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.67)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.67)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.67)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded -0.18)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
45% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
45% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
45% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
45% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
45% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.21)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.39)
35% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.39)
35% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.39)
35% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.39)
35% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.39)
35% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.39)
35% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.39)
35% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.39)
35% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.39)
35% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.39)
35% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.39)
35% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.39)
35% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.39)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.85)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.38)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.33)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.33)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.33)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.33)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.33)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.33)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.33)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.33)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.33)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.55)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.11)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.99)
5% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.99)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 3
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 2.20)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.14)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.67)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.67)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.67)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.15)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.79)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.89)
68% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.89)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.36)
64% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.36)
64% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.36)
64% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.36)
64% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.36)
64% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.36)
64% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.36)
64% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.36)
64% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.36)
64% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.36)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 1.69)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.84)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.84)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.84)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.84)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.84)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.84)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.84)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.84)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.84)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.84)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.84)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.84)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.84)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.84)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.82)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.15)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.50)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 0.86)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 0.86)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 0.86)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 0.86)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 0.86)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.11)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.77)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 0.77)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.28)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.28)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.39)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.60)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.08)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.34)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.54)
4% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.54)
4% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.54)
4% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.54)
4% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.54)
4% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.54)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.95)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 4
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.44)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.03)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.61)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.94)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.26)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.16)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.22)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.22)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.22)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.66)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.51)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.53)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.90)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.81)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 2.73)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.79)
44% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.79)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 2.19)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.09)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.62)
32% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.62)
32% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.62)
32% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.62)
32% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.62)
32% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.62)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.11)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.06)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.96)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.64)
16% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.64)
16% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.64)
16% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.64)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.88)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.18)
8% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.18)
8% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.18)
8% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.18)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.30)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.95)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 5
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 2.84)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.67)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.29)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.38)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.07)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.07)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.07)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.07)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.07)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.07)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.07)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.07)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.73)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.43)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.59)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.32)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.48)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.94)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.40)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.41)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.28)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.09)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.06)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 1.09)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.43)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.88)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 6
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.80)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.30)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.84)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.84)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.84)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.84)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.84)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.84)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.84)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.84)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.84)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.84)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.84)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.84)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.84)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.77)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.44)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.58)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.58)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.58)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.58)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.58)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.58)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.58)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.58)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.58)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.58)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.58)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.81)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.40)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.72)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.72)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.72)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.72)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.72)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.72)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.72)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded -0.15)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.90)
48% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.90)
48% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.90)
48% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.90)
48% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.90)
48% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.90)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.98)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded -0.13)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded -0.15)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.53)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 2.04)
28% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 2.04)
28% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 2.04)
28% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 2.04)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 1.44)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.38)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.54)
16% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.54)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
12% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
12% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
8% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded -0.37)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.42)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 7
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.47)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.01)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.12)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.05)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.97)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 0.99)
70% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 0.99)
70% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 0.99)
70% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 0.99)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 1.16)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.09)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.29)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.29)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.29)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.29)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.29)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.05)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.79)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.08)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.55)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.27)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.91)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.07)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.18)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.46)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.42)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 8
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.59)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.43)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.28)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.16)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.18)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.72)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.55)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.79)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.54)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.31)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.65)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.52)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.61)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.61)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.61)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.61)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.61)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.61)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.61)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.61)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.61)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.66)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
32% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
32% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
32% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
32% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
32% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
32% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
32% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
32% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
32% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
32% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.39)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.39)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.39)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.39)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.39)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.39)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.39)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.39)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.39)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.23)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
12% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
12% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
12% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
12% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
12% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
12% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
12% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
12% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
12% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
12% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
12% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.35)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.65)
4% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.65)
4% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.65)
4% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.65)
4% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.65)
4% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.65)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded -0.03)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 9
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.33)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.12)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.04)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.42)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.71)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.48)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.30)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.57)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.36)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.25)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.33)
52% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.33)
52% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.33)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.52)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.13)
44% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.13)
44% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.13)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.89)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.31)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.32)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.70)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded -0.40)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.36)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.68)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.02)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.27)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.43)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 10
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.01)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.20)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.24)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.04)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded -0.01)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.71)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.24)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.89)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.04)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.04)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.04)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.04)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.04)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.04)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.04)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.04)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.04)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.04)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.04)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.85)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.88)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 2.68)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 0.91)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 11
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.97)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.19)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.67)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.62)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.20)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.20)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.20)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.20)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.20)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.20)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.20)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.20)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.20)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.20)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.20)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.20)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.20)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.96)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.21)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.07)
50% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.07)
50% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.07)
50% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.07)
50% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.07)
50% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.07)
50% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.07)
50% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.07)
50% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.07)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.76)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.65)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 0.75)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.30)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.60)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.05)
15% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.05)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.95)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.81)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.49)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 12
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.78)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.33)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.83)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.59)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.18)
76% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.18)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.03)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.13)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.16)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.89)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.11)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.11)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.11)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.14)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.67)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.76)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.61)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.83)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.67)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.35)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.67)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.08)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.22)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 1.13)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.31)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.75)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 13
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.08)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.63)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.02)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.68)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.95)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.69)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.75)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.70)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.28)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.62)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded -0.24)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.25)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.99)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.58)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.63)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.24)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.71)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded -0.69)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 14
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.81)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.58)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.46)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.66)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.89)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.25)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 2.65)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.04)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.42)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.06)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.48)
63% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.48)
63% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.48)
63% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.48)
63% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.48)
63% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.48)
63% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.48)
63% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.48)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.13)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.51)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.91)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 1.59)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 1.39)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.22)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.72)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.72)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.72)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.72)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.85)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.27)
33% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.27)
33% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.27)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.35)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.42)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.13)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.63)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.46)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded -0.40)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.70)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.03)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.87)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 15
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.55)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.89)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.52)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.07)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.51)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.96)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.28)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.46)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.78)
69% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.78)
69% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.78)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.26)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.39)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.73)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.73)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.73)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.47)
57% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.47)
57% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.47)
57% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.47)
57% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.47)
57% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.47)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.75)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.04)
51% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.04)
51% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.04)
51% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.04)
51% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.04)
51% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.04)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.83)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.34)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.10)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.32)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.23)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.00)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.55)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.96)
29% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.96)
29% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.96)
29% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.96)
29% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.96)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.34)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.39)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.81)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.14)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.59)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 0.94)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.52)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.34)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.97)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.97)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.97)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.97)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.97)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.97)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.97)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.97)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.97)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.97)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.93)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 16
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.64)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.77)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.02)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.06)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.06)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.06)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.06)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.06)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.06)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.06)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.06)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.06)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.06)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.82)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 2.61)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.61)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.91)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.08)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.18)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.08)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.59)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 17
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.66)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.39)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.43)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.69)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.45)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.87)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.11)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 2.58)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.69)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.14)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.14)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.14)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.14)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.14)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.14)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.14)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.14)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.14)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.14)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.14)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.14)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.14)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.14)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.14)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.98)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.71)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded -0.07)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.90)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.90)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.57)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.67)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.06)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.30)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.82)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded -0.49)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.80)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.69)
4% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.69)
4% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.69)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.57)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 18
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.23)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.45)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.70)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.37)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.85)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.35)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.17)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.25)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 1.57)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.63)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.07)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded -0.03)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.28)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.35)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded -0.25)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.49)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded -0.57)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.56)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 19
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.20)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.66)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.84)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.63)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.92)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.06)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.06)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.54)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.73)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.26)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.91)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.68)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.12)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.05)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.88)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.11)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.79)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.84)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.20)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.04)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded -0.00)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.35)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.43)
23% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.43)
23% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.43)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.04)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.72)
17% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.72)
17% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.72)
17% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.72)
17% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.72)
17% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.72)
17% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.72)
17% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.72)
17% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.72)
17% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.72)
17% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.72)
17% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.72)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.13)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.08)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.54)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.54)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.54)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.54)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.54)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.54)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.54)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.54)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.54)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.54)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.54)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.54)
3% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.54)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.31)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 20
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.51)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.37)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.61)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.61)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.61)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.61)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.61)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.61)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.61)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.61)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.61)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.61)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.61)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.61)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.61)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.61)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.95)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.70)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.70)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.70)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.70)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.70)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.70)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.05)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.11)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.11)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.11)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.11)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.11)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.11)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.11)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.11)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.11)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.87)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.92)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.88)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.67)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded -0.17)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.30)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.68)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.23)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.93)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.27)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 1.52)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.59)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.59)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.59)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.59)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.59)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.59)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.59)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.59)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.59)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.59)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.59)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.59)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.59)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.59)
24% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.59)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.25)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.43)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.62)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.74)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.77)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 1
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.83)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.97)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.61)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.80)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.80)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.80)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.80)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.80)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.80)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.80)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.80)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.80)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.80)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.80)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.80)
75% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.80)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 1.05)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.88)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.30)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.22)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.58)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.52)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.42)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 2.20)
35% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 2.20)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.69)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.19)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.34)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.13)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 0.60)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Testing trial 2
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.78)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.75)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.51)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.52)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.81)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 2.34)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.60)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.91)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.23)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.60)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.33)
52% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.33)
52% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.33)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.59)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.66)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.34)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.43)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.34)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.17)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.04)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.04)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.04)
20% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.04)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.27)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.20)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.53)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.76)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 3
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 1.43)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.83)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 1.81)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.61)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.64)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.63)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.25)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.90)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.02)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.15)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.49)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.02)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.72)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 1.52)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.27)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.14)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.71)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.58)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.94)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 2.27)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.83)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.60)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 4
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.79)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.05)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.87)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.51)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.51)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.63)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 1.00)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 1.52)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.61)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.39)
50% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.39)
50% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.39)
50% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.39)
50% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.39)
50% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.39)
50% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.39)
50% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.39)
50% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.39)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.12)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded -0.05)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded -0.05)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded -0.05)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded -0.05)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded -0.05)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.99)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.28)
25% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.28)
25% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.28)
25% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.28)
25% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.28)
25% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.28)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.22)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.20)
15% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.20)
15% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.20)
15% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.20)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.36)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.70)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 5
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.46)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.30)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.17)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.95)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 2.51)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.28)
70% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.28)
70% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.28)
70% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.28)
70% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.28)
70% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.28)
70% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.28)
70% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.28)
70% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.28)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.29)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.49)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.13)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.48)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.88)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.74)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.26)
35% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.26)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.66)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.82)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.60)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.11)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint forward. (rewarded 1.73)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.62)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.57)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 6
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.33)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.80)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.80)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.80)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.80)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.80)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.80)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.80)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.80)
88% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.80)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.01)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.06)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.50)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 1.29)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.25)
68% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.25)
68% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.25)
68% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.25)
68% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.25)
68% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.25)
68% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.25)
68% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.25)
68% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.25)
68% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.25)
68% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.25)
68% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.25)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.23)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.71)
60% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.71)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.54)
56% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.54)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.85)
52% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.85)
52% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.85)
52% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.85)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.64)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded -0.24)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.52)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.66)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.37)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded -0.09)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.19)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.27)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.18)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.05)
8% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.05)
8% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.05)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.34)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.02)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 7
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.00)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.47)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.90)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 0.47)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.09)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.35)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.61)
72% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.61)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 1.36)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.98)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.67)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 1.43)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.57)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.31)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.16)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.44)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.31)
36% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -9.31)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.02)
32% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.02)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.04)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.73)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.37)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.40)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded -0.58)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
8% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
8% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
8% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded -0.71)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded -0.03)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 8
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 1.17)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.27)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.62)
85% of time remaining to reach destination.
!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.62)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.20)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.20)
80% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.20)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.00)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.99)
70% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.99)
70% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.99)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 1.32)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 1.32)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove left instead of forward. (rewarded 1.32)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.42)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 1.58)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 1.17)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.75)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.56)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.73)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
30% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
30% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.16)
25% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.16)
25% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.16)
25% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.16)
25% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.16)
25% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.16)
25% of time remaining to reach destination.
!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.16)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.38)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove left instead of right. (rewarded 0.36)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.08)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.41)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.27)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 9
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.37)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.20)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.94)
85% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.94)
85% of time remaining to reach destination.
!! Agent state not been updated!
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.94)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 1.81)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 1.66)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.13)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.38)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.38)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.38)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.38)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.38)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.38)
65% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.38)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.98)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.24)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.24)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.24)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.24)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.24)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.24)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.24)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.24)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.24)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.24)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.24)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.24)
55% of time remaining to reach destination.
!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.24)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -10.90)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.93)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.86)
40% of time remaining to reach destination.
!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.86)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.95)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.15)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded -0.17)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.72)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving forward through a red light. (rewarded -10.46)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.27)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 10
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint right. (rewarded 2.93)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of forward. (rewarded 0.54)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 1.54)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.19)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.17)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.24)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

!! Agent state not been updated!
Agent followed the waypoint left. (rewarded 2.25)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.25)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 1.97)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 2.64)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -5.33)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of left. (rewarded 0.48)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.08)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

!! Agent state not been updated!
Agent drove right instead of left. (rewarded 0.41)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded 0.89)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

!! Agent state not been updated!
Agent properly idled at a red light. (rewarded -0.27)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.83)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

!! Agent state not been updated!
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.67)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded -0.26)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.41)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded -0.21)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.86)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

!! Agent state not been updated!
Agent drove forward instead of right. (rewarded 0.12)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

!! Agent state not been updated!
Agent attempted driving left through a red light. (rewarded -9.65)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

Simulation ended. . . 
In [22]:
# Load the 'sim_no-learning' log file from the initial simulation results
vs.plot_trials('./sim_no-learning.csv')

Question 3

Using the visualization above that was produced from your initial simulation, provide an analysis and make several observations about the driving agent. Be sure that you are making at least one observation about each panel present in the visualization. Some things you could consider:

  • How frequently is the driving agent making bad decisions? How many of those bad decisions cause accidents?
  • Given that the agent is driving randomly, does the rate of reliability make sense?
  • What kind of rewards is the agent receiving for its actions? Do the rewards suggest it has been penalized heavily?
  • As the number of trials increases, does the outcome of results change significantly?
  • Would this Smartcab be considered safe and/or reliable for its passengers? Why or why not?

Answer:

  • 1.The driving agent was making bad decision almost 45% of the time at the beginning (44.25%) resulting in major accidents around 4-4.5% of the time, but this gradually dropped to about 38% resulting in less than 4% major accidents.When the rate of bad decision increases to 43% which leads to increase in major accident to 10% of time.
  • 2.Given that the agent is not learning and driving randomly , it is late for atleast 80% of the time , this rate make sense.
  • 3.Intially reward starts at -4.7 and then it increases to -3.7 this is due to rate of making bad decision decreased and then again it decrease to -5 tis is due to increase in bad decision and major accident . So the rate of change in Rewards is related to rate of bad decisions and major accidents.
    1. The outcome of relaiability remains same as the initial but in between there was change that is due to the decrease in bad decisions and major accident . with respect to The rate of bad decisons there was 10% improvement and with respect to major accident infact the rate as increased compard to initial rate.
  • %. No this smart cab is not safe because its actions being random, results in accidents, both major and minor, 7% of the time with major and minor triffic violations a little less than 5% of the time even at the end of 20 trials. It is not reliable because it is also consistantly late 80-90% of the time.

Inform the Driving Agent

The second step to creating an optimized Q-learning driving agent is defining a set of states that the agent can occupy in the environment. Depending on the input, sensory data, and additional variables available to the driving agent, a set of states can be defined for the agent so that it can eventually learn what action it should take when occupying a state. The condition of 'if state then action' for each state is called a policy, and is ultimately what the driving agent is expected to learn. Without defining states, the driving agent would never understand which action is most optimal -- or even what environmental variables and conditions it cares about!

Identify States

Inspecting the 'build_state()' agent function shows that the driving agent is given the following data from the environment:

  • 'waypoint', which is the direction the Smartcab should drive leading to the destination, relative to the Smartcab's heading.
  • 'inputs', which is the sensor data from the Smartcab. It includes
    • 'light', the color of the light.
    • 'left', the intended direction of travel for a vehicle to the Smartcab's left. Returns None if no vehicle is present.
    • 'right', the intended direction of travel for a vehicle to the Smartcab's right. Returns None if no vehicle is present.
    • 'oncoming', the intended direction of travel for a vehicle across the intersection from the Smartcab. Returns None if no vehicle is present.
  • 'deadline', which is the number of actions remaining for the Smartcab to reach the destination before running out of time.
In [32]:
%run smartcab/agent.py
/-------------------------
| Training trial 1
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.04)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent drove forward instead of left. (rewarded 0.21)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.51)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.58)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.58)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded -0.01)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.55)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.18)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.16)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.78)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.53)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.92)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.92)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.92)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.65)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.85)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.40)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.22)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.23)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.79)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.50)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.96)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.99)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 2
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.54)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 1.74)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.50)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.50)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.41)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.76)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.94)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.58)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 2.78)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 2.78)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.98)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.10)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.35)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.64)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
49% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
49% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
49% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
49% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
49% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.73)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.56)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.16)
37% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.16)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.56)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.53)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.94)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.02)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.02)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.02)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.02)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.02)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.20)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.18)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.61)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.34)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.47)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.20)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.13)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.88)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 3
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.13)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.78)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.40)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.42)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 1.86)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.11)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.85)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.85)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.85)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.85)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.77)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.14)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.05)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 1.56)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.26)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 1.63)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.55)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.54)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.80)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.52)
37% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.52)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.03)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.91)
27% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 4
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.35)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.51)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.63)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.42)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.29)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.34)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.34)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.34)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.34)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.34)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.95)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.37)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.46)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.23)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.23)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.23)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.23)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.23)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.23)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.25)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.78)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.75)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.75)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.75)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.72)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded 0.11)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.07)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 5
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.86)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.49)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.88)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.24)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove right instead of left. (rewarded 1.90)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.53)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.63)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.93)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.97)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.51)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.94)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.24)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.02)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded -0.32)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded -0.38)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.66)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 6
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.41)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.95)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.11)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.34)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.55)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.69)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.69)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.69)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.69)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.69)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.69)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.69)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.69)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.06)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.06)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.75)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.75)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.08)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.13)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.01)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded -0.11)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded -0.11)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded -0.11)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded -0.11)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded -0.11)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.91)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.25)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 1.41)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 1.41)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.17)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.54)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.45)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.48)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded -0.41)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.07)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.14)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 7
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.50)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.57)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.81)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.82)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.82)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.82)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.82)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.82)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.82)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.82)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.49)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.01)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.61)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.02)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.06)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.97)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.24)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.85)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.06)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.15)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.15)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.15)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.15)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.15)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.15)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.15)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.19)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.27)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 8
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove forward instead of right. (rewarded 1.88)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.11)
93% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.11)
93% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.11)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.23)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 0.20)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.43)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.35)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.58)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.45)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.55)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.39)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.07)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.98)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.42)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.92)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.03)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.73)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.00)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.95)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.21)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.66)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.16)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.74)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.74)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.74)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.74)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.23)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 1.27)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.33)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.99)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.44)
3% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.44)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.23)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 9
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.00)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.60)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.10)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.08)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 1.05)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.72)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.12)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.20)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.46)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.20)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 10
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.21)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.18)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.22)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.62)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.97)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.63)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.60)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.60)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.60)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.60)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.80)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.33)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.68)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.42)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 11
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove right instead of forward. (rewarded 1.35)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.79)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.39)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 1.36)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.22)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.65)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 0.45)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.71)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.71)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.71)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.71)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.71)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 0.87)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.55)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.38)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.89)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.89)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.89)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.89)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.89)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.06)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.06)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.06)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.06)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.06)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.06)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.06)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.06)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.06)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.72)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.43)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.99)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 12
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.59)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.32)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.85)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.85)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 1.08)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.21)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.54)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.54)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.54)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 0.97)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.75)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.75)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.75)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.75)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.75)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.75)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.75)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.75)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.75)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -6.00)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.71)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.71)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.78)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.39)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.17)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.43)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.69)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.59)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.59)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.59)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.96)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.46)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 13
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent drove forward instead of right. (rewarded 1.09)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.59)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.64)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.84)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.44)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.75)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.75)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.75)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.37)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 0.35)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 0.96)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.74)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.74)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.74)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.74)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.74)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.74)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.74)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.74)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.52)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.11)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.13)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.68)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.43)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.82)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.57)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.60)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.20)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.45)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.39)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
31% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
31% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
31% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
31% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
31% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
31% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.08)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.22)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.81)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.49)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.74)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.46)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.14)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.38)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.59)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent drove right instead of left. (rewarded -0.19)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 14
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.86)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.40)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.53)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.51)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.43)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.88)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.88)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.88)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.88)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.88)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.02)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.78)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.34)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.34)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.34)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.34)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.31)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.37)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove left instead of forward. (rewarded 0.15)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.46)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.46)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.46)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.29)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.74)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.94)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.11)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 0.40)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 0.57)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 15
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.83)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.10)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.77)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.71)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.43)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.56)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.86)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.86)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.86)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.86)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent drove right instead of left. (rewarded 1.76)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.01)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.73)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.73)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.73)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.92)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.34)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.34)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.34)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.02)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.51)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.28)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.34)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.01)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.30)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.30)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.30)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.30)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.53)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 16
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.53)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.96)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.92)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.79)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.14)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.25)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded -0.10)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.73)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.64)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.46)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 1.60)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.44)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.34)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.33)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.33)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.33)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.33)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.33)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.17)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 2.40)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 2.40)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.07)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded -0.05)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.76)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.91)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.67)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 17
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.64)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.72)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.46)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.79)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.55)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded -0.01)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded -0.01)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded -0.01)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.50)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.50)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.50)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.50)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.82)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.49)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.84)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.75)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.75)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.90)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.08)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.51)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.34)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.34)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.34)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.34)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.34)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.34)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.34)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.34)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 18
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.38)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.88)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.25)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.25)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.25)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.25)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.25)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove left instead of right. (rewarded 1.28)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove forward instead of left. (rewarded 0.48)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.14)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.06)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.88)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.80)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.86)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.08)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.08)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.08)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.88)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent drove right instead of left. (rewarded 1.22)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 1.26)
28% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 1.26)
28% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 1.26)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.54)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.56)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.83)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.71)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.60)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 19
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.07)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -11.00)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.87)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.89)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.69)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.86)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.69)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.22)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.66)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.01)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.01)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.01)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.50)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.59)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded 1.35)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.01)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.15)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.49)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.90)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 20
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.51)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.85)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.85)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.04)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.70)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.63)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.86)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.57)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.56)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.51)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.43)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.45)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.45)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.45)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.02)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.13)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.13)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.53)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.53)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.53)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.53)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.85)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.10)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 1
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.02)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.05)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.66)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 1.29)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.17)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.52)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.93)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.93)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.93)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.84)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.88)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.88)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.73)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.48)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.84)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.15)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.15)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.15)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.15)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.06)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.67)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.66)
27% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.66)
27% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.66)
27% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.66)
27% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.66)
27% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.66)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.83)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.52)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.47)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 2
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.34)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.05)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.28)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.28)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.28)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.28)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.28)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.28)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.28)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.28)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.78)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.54)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.51)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.73)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.73)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.73)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.73)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.73)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.73)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.73)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.73)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.73)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.34)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.01)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.39)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.67)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.36)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.82)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.01)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.48)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.37)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 3
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.85)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.15)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.90)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.72)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.78)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.79)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 1.20)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.56)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.79)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.68)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.78)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove forward instead of right. (rewarded -0.08)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.50)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.52)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.42)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.18)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.18)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.34)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent drove forward instead of right. (rewarded 1.47)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.26)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.26)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.26)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.26)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.26)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.58)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.65)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.47)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 4
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.65)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.87)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.88)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.49)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.30)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.94)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.94)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.19)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.19)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.19)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.95)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.16)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.13)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.94)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.24)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.60)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.34)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.86)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.48)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.89)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.37)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 5
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.86)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.54)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.18)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.40)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.40)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.40)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.40)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.40)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.40)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.40)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.40)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent drove left instead of right. (rewarded 0.59)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.01)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.19)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.19)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 0.89)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.35)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.89)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.89)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.89)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.89)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.89)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.27)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.26)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 6
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.35)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.76)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.54)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.79)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.02)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.49)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.49)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.49)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.49)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.49)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.49)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.49)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.49)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.70)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.97)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.18)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.92)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.86)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.32)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.63)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.12)
36% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Testing trial 7
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.02)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.50)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.69)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.69)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.69)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.78)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.89)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 0.88)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded 1.03)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.78)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.08)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.34)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.62)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.62)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.62)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.40)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.93)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.93)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.93)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.93)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.50)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.31)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded -0.74)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent drove right instead of left. (rewarded -0.54)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 8
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.16)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.62)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.45)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.82)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.82)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.82)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.82)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.25)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.55)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.76)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.14)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.61)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.65)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.27)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.82)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.19)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.88)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.80)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.80)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.84)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.62)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.40)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.69)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.44)
27% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.44)
27% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.44)
27% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.44)
27% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.44)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.91)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.90)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.35)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.58)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.58)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.87)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.28)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 9
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.46)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.45)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
91% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.96)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.96)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.96)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.96)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.96)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.71)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.44)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.89)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.71)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.47)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.88)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove right instead of left. (rewarded 0.85)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.24)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.52)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.20)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.20)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.58)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent attempted driving left through a red light. (rewarded -9.37)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.07)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.08)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 1.90)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 1.90)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.02)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.02)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.02)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.26)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.51)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.45)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.52)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent drove right instead of left. (rewarded 0.41)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.70)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.70)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.70)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.65)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.44)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.48)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded -0.28)
6% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded -0.28)
6% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded -0.28)
6% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded -0.28)
6% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded -0.28)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded -0.45)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded -0.55)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 10
\-------------------------

Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.
Simulating trial. . . 
Agent not set to learn.

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent drove forward instead of left. (rewarded 1.88)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 2.42)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.36)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.02)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.02)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.02)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.53)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.53)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.62)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.58)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.64)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.74)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.74)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.28)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.41)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.67)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.11)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded -0.56)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.53)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.23)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.23)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.50)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

Simulation ended. . . 

Question 4

Which features available to the agent are most relevant for learning both safety and efficiency? Why are these features appropriate for modeling the Smartcab in the environment? If you did not choose some features, why are those features not appropriate? Please note that whatever features you eventually choose for your agent's state, must be argued for here. That is: your code in agent.py should reflect the features chosen in this answer.

NOTE: You are not allowed to engineer new features for the smartcab.

Answers

Answer:

  • The most relevant feature for learning both safety and efficiency are
    1. Waypoint: It tells the agent about the direction to take to reach intended destination. Taking the correct route to reach destination will increase the efficiency of the smartcab which inturn increase the reliability.
    1. inputs is the sensor data from the agent in the environment a.light: the color of the signal, which makes the agent to learn about the traffic signal and if it stays idle when the signal is green then it will be penalized and if it goes forward when the signal is red then it will be penalized high which are major traffic voilation.Based on the reward, it will learn the action to be followed to drive safely. b.oncoming: smart cab must know any vehicle in oncoming direction in order to take safe left turn . c.Left: It tells about the presence of vehicle on the left side of smart cab and its value tells that vehicle intnded action. If smartcab is in intersection and the signal is red/green and if there is a vehicle in the left and its intended action is forward then smartcab taking left action will result in major accident. d.input:right - If the smart cab is in intersection and if there is vehicle in right and its intended action to go forward then smart cab performing right action will lead to accidents. So these features are relevant for safety and efficiency.
    2. Not relevant features:
  • deadline not as relevent because I do not believe the Smartcab agent needs to know the deadline, but can learn about it in the reward it receives as it builds its Q table. Setting it this way may help improve on its reliability by causing minor safety violations to major accidents. This is something to avoid, but I think we want the agent to learn it on its own. Plus, doing it this way significantly reduces the Q table size and makes the learning algorthm more efficient.

Define a State Space

When defining a set of states that the agent can occupy, it is necessary to consider the size of the state space. That is to say, if you expect the driving agent to learn a policy for each state, you would need to have an optimal action for every state the agent can occupy. If the number of all possible states is very large, it might be the case that the driving agent never learns what to do in some states, which can lead to uninformed decisions. For example, consider a case where the following features are used to define the state of the Smartcab:

('is_raining', 'is_foggy', 'is_red_light', 'turn_left', 'no_traffic', 'previous_turn_left', 'time_of_day').

How frequently would the agent occupy a state like (False, True, True, True, False, False, '3AM')? Without a near-infinite amount of time for training, it's doubtful the agent would ever learn the proper action!

Question 5

If a state is defined using the features you've selected from Question 4, what would be the size of the state space? Given what you know about the environment and how it is simulated, do you think the driving agent could learn a policy for each possible state within a reasonable number of training trials?
Hint: Consider the combinations of features to calculate the total number of states!

Answer:

  • First, we must consider each feature, and the number of states each one has:
  • 1.waypoint - [left, right, forward] : 3
  • 2.inputs-light - [red, green] : 2
  • 3.inputs-left - [left, right, forward, None] : 4
  • 4.inputs-right - [left, right, forward, None] : 4
  • 5.inputs-oncoming - [left, right, forward, None] : 4
  • total number of states: 3 x 2 x 4 x 4 x 4 = 384 states.
  • possible Action - [left, right, forward, None]
  • If we add possible action then 3 x 2 x 4 x 4 x 4 x 4 = 1536 possible combination.
  • I believe the driving agent could learn a policy for each possible state within a reasonable number of training trials.

Update the Driving Agent State

For your second implementation, navigate to the 'build_state()' agent function. With the justification you've provided in Question 4, you will now set the 'state' variable to a tuple of all the features necessary for Q-Learning. Confirm your driving agent is updating its state by running the agent file and simulation briefly and note whether the state is displaying. If the visual simulation is used, confirm that the updated state corresponds with what is seen in the simulation.

Note: Remember to reset simulation flags to their default setting when making this observation!


Implement a Q-Learning Driving Agent

The third step to creating an optimized Q-Learning agent is to begin implementing the functionality of Q-Learning itself. The concept of Q-Learning is fairly straightforward: For every state the agent visits, create an entry in the Q-table for all state-action pairs available. Then, when the agent encounters a state and performs an action, update the Q-value associated with that state-action pair based on the reward received and the iterative update rule implemented. Of course, additional benefits come from Q-Learning, such that we can have the agent choose the best action for each state based on the Q-values of each state-action pair possible. For this project, you will be implementing a decaying, $\epsilon$-greedy Q-learning algorithm with no discount factor. Follow the implementation instructions under each TODO in the agent functions.

Note that the agent attribute self.Q is a dictionary: This is how the Q-table will be formed. Each state will be a key of the self.Q dictionary, and each value will then be another dictionary that holds the action and Q-value. Here is an example:

{ 'state-1': { 
    'action-1' : Qvalue-1,
    'action-2' : Qvalue-2,
     ...
   },
  'state-2': {
    'action-1' : Qvalue-1,
     ...
   },
   ...
}

Furthermore, note that you are expected to use a decaying $\epsilon$ (exploration) factor. Hence, as the number of trials increases, $\epsilon$ should decrease towards 0. This is because the agent is expected to learn from its behavior and begin acting on its learned behavior. Additionally, The agent will be tested on what it has learned after $\epsilon$ has passed a certain threshold (the default threshold is 0.05). For the initial Q-Learning implementation, you will be implementing a linear decaying function for $\epsilon$.

Q-Learning Simulation Results

To obtain results from the initial Q-Learning implementation, you will need to adjust the following flags and setup:

  • 'enforce_deadline' - Set this to True to force the driving agent to capture whether it reaches the destination in time.
  • 'update_delay' - Set this to a small value (such as 0.01) to reduce the time between steps in each trial.
  • 'log_metrics' - Set this to True to log the simluation results as a .csv file and the Q-table as a .txt file in /logs/.
  • 'n_test' - Set this to '10' to perform 10 testing trials.
  • 'learning' - Set this to 'True' to tell the driving agent to use your Q-Learning implementation.

In addition, use the following decay function for $\epsilon$:

$$ \epsilon_{t+1} = \epsilon_{t} - 0.05, \hspace{10px}\textrm{for trial number } t$$

If you have difficulty getting your implementation to work, try setting the 'verbose' flag to True to help debug. Flags that have been set here should be returned to their default setting when debugging. It is important that you understand what each flag does and how it affects the simulation!

Once you have successfully completed the initial Q-Learning simulation, run the code cell below to visualize the results. Note that log files are overwritten when identical simulations are run, so be careful with what log file is being loaded!

In [43]:
%run smartcab/agent.py
/-------------------------
| Training trial 1
\-------------------------

Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9500; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.97)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.14)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.62)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.66)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.57)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.08)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.08)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.08)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.08)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.08)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.08)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.08)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.17)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.09)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.35)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.79)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 0.28)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent drove left instead of forward. (rewarded 0.74)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded -0.39)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.03)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.16)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 2
\-------------------------

Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.9000; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'right')
Agent followed the waypoint right. (rewarded 1.88)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.77)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.85)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.70)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.53)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.35)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.73)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent drove forward instead of right. (rewarded 0.75)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.34)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.34)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.98)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.58)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.01)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.01)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.01)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.01)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.01)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.43)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 0.78)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.56)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.48)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded -0.72)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 3
\-------------------------

Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8500; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.62)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.81)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.09)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.27)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 1.29)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.70)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.68)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.33)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.85)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.17)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent drove right instead of forward. (rewarded -0.38)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.10)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.64)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.61)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.88)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 4
\-------------------------

Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.8000; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.85)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.04)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.92)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.70)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove left instead of right. (rewarded 1.37)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.62)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.86)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.27)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.03)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.03)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.03)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.03)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.03)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.03)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.03)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.39)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.81)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.18)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.18)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.01)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.01)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.01)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.01)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.21)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.99)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.84)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.45)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.65)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.70)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.40)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.54)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.47)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.60)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 5
\-------------------------

Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7500; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.74)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.14)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.01)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.34)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.36)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.36)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.60)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.60)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.60)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.73)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.61)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.61)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.61)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.06)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.06)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.99)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.47)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.77)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.47)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.32)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.22)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.89)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 6
\-------------------------

Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.7000; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.58)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.38)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.54)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.87)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.67)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 1.15)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.06)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.06)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.06)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.42)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.42)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.63)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove left instead of right. (rewarded 0.07)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.66)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.66)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.66)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.66)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.45)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.23)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.23)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.23)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.23)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.04)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.33)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 1.36)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 0.65)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 0.65)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.85)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.85)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.85)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.00)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.00)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.00)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded -0.48)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 7
\-------------------------

Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6500; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 1.01)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.64)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.45)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.62)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.08)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.26)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.49)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.49)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.49)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.49)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.49)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.49)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.49)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.73)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.76)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.27)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.17)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'right')
Agent properly idled at a red light. (rewarded 1.23)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', 'right')
Agent properly idled at a red light. (rewarded 1.23)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', 'right')
Agent properly idled at a red light. (rewarded 1.23)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.20)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.65)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.65)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.65)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.65)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.65)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.65)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.52)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.24)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.75)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.09)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.09)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.09)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.09)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.09)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.04)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.53)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -10.71)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.61)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 8
\-------------------------

Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.6000; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.20)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.50)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.11)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.00)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.70)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.70)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.70)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.70)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.44)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.40)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.40)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.40)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.40)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.40)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.40)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.31)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.59)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.05)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.78)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.69)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.00)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.31)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.31)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.31)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.31)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.31)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.31)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.91)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.72)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded -0.11)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 9
\-------------------------

Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5500; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.99)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.63)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.06)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.81)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.12)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.12)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.12)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.12)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.12)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.12)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.12)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.36)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.14)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.60)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.28)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.10)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.10)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.10)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.10)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.10)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.10)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.10)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 0.41)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.12)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.12)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.12)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.29)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.79)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent drove right instead of left. (rewarded -0.20)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.68)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.23)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded -0.35)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.38)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.57)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.76)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 10
\-------------------------

Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.5000; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.57)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.03)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.55)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.57)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.41)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 11
\-------------------------

Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4500; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.58)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.34)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.08)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.87)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.92)
73% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 12
\-------------------------

Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.4000; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 1.26)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove left instead of right. (rewarded 0.16)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.57)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.13)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.97)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.85)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.80)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.30)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.89)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.89)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.64)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.83)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.04)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.07)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.72)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.37)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.40)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 13
\-------------------------

Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3500; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.40)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.06)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.07)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.33)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.91)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.27)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.90)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.78)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.24)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.60)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.37)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 1.05)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.17)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.17)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.17)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.17)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.68)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.25)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.02)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 14
\-------------------------

Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.80)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.89)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.30)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.30)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.28)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.99)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.36)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.51)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.80)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.18)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.12)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.54)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.55)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.83)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.44)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 1.13)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.12)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.19)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.19)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.19)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.19)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.19)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.19)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.33)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 15
\-------------------------

Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2500; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.67)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.74)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.88)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.40)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.43)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.22)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.66)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.49)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.23)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 2.67)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.20)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.20)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.20)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.20)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.20)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.39)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.68)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.55)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.55)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.55)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.55)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.05)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.54)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.09)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.09)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.81)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.22)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.36)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.16)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.58)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 16
\-------------------------

Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.2000; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded 1.29)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent drove right instead of left. (rewarded 1.50)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.75)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.75)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.75)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.75)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.14)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.15)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.27)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.27)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.27)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.55)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.54)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.76)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.02)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.02)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.02)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.27)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.27)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.27)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.59)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 17
\-------------------------

Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1500; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.01)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.05)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.04)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.05)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.94)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.10)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.14)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.25)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.97)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.97)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.97)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.97)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.97)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 1.45)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.19)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.19)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.19)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.19)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.19)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.19)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.71)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.37)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.37)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.37)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.37)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.37)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.70)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.67)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.68)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.50)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.50)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.50)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.50)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.50)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.50)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.26)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 0.98)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.61)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.27)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.81)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.58)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 18
\-------------------------

Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.1000; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.73)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded -0.01)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.06)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.06)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.06)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.06)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.79)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.79)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.79)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.79)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.79)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.45)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.45)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.45)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.45)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.17)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.37)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.75)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.26)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.73)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.22)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.60)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.60)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.60)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.60)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.03)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.00)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', 'forward')
Agent followed the waypoint forward. (rewarded 0.90)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.27)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.27)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.27)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.27)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.47)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.65)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.42)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 0.11)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.38)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.92)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.65)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.65)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.65)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.65)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.65)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.89)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 19
\-------------------------

Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.69)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.90)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.20)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded 1.12)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.40)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.18)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.19)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.93)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.16)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.79)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.08)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 20
\-------------------------

Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . . 
epsilon = -0.0000; alpha = 0.5000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'right')
Agent drove right instead of left. (rewarded 1.98)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.31)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.71)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.71)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.71)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.29)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 1.52)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.21)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.54)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.80)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.16)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.31)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 1.07)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 1.07)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 1.07)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.94)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.11)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.46)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.75)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.75)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 1.10)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded -0.32)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.66)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.75)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.64)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.47)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.77)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.70)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.70)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.70)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.70)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.50)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.09)
3% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Testing trial 1
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.05)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.83)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.49)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.65)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.65)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.12)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.22)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.18)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.71)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.09)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.47)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.47)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.47)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.47)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.47)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.49)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.49)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.49)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.49)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.49)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.23)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.42)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.17)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.17)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.17)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.17)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.17)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.17)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.70)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent drove left instead of right. (rewarded 0.10)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent drove left instead of right. (rewarded 0.10)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent drove left instead of right. (rewarded 0.10)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent drove left instead of right. (rewarded 0.10)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent drove left instead of right. (rewarded 0.10)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent drove left instead of right. (rewarded 0.10)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.12)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 2
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.49)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.28)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.92)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.55)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.85)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.78)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Testing trial 3
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove right instead of left. (rewarded 0.26)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.86)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.64)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.66)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.93)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.83)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.73)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.23)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.55)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.08)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.56)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.56)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.70)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.70)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.62)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.33)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.33)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.60)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.78)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.96)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.64)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.16)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.91)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.00)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.00)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.00)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.35)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 4
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.83)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.43)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.03)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.03)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.03)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.03)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.31)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.86)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.12)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.52)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.84)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.85)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.96)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.46)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.31)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.69)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.12)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.49)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.44)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.50)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.50)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.50)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.98)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.05)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.68)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded -0.27)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.33)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.33)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', 'forward')
Agent followed the waypoint forward. (rewarded 0.46)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 5
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.93)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.26)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.26)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.26)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.45)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.13)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.13)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.57)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.57)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.57)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.57)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.67)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.83)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.83)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.21)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.11)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 0.90)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent drove right instead of forward. (rewarded 1.36)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.12)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded -0.15)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.56)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded -0.31)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.91)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.20)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.20)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.20)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.20)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.20)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.20)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.20)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.38)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.36)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 6
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.44)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.76)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove left instead of right. (rewarded 0.08)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.79)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.10)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.67)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.17)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.35)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Testing trial 7
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.15)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.15)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.15)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.78)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.63)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.53)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.67)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.91)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.01)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.43)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.04)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.22)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.65)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.11)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.94)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.79)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.43)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 8
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.64)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.42)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.64)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.66)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.49)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.06)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.06)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.06)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.88)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.04)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.20)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.20)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.76)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.18)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.71)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.13)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.01)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.71)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.17)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.17)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.04)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent drove forward instead of right. (rewarded -0.02)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.80)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.03)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.70)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 9
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.31)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.90)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.54)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.56)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.88)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 1.55)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.05)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.16)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.93)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'right', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.70)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded -0.06)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.61)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.57)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.35)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.53)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded -0.50)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.10)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.60)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 10
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.81)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.74)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.30)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.13)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.48)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded 1.30)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded 1.30)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded 1.30)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.67)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.83)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.82)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.40)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.78)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.19)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.13)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

Simulation ended. . . 
<matplotlib.figure.Figure at 0x935bb70>
In [44]:
# Load the 'sim_default-learning' file from the default Q-Learning simulation
vs.plot_trials('sim_default-learning.csv')

Question 6

Using the visualization above that was produced from your default Q-Learning simulation, provide an analysis and make observations about the driving agent like in Question 3. Note that the simulation should have also produced the Q-table in a text file which can help you make observations about the agent's learning. Some additional things you could consider:

  • Are there any observations that are similar between the basic driving agent and the default Q-Learning agent?
  • Approximately how many training trials did the driving agent require before testing? Does that number make sense given the epsilon-tolerance?
  • Is the decaying function you implemented for $\epsilon$ (the exploration factor) accurately represented in the parameters panel?
  • As the number of training trials increased, did the number of bad actions decrease? Did the average reward increase?
  • How does the safety and reliability rating compare to the initial driving agent?

Answer:

    1. When comparing with basic driving agent, the overall rating is the same - achieving an F for safety and an F for reliability.
  • But there is reduction in bad actions:
  • Major Violations have dropped from 19% to 7%
  • Minor Violations have dropped from 7% to 3%
  • Overall, bad actions have dropped by 34% to 19%
  • But
  • Major Accidents and Minor Accidents starting and ending of trial remains the same but in between there was decrese in the acidents.
  • And Rate of reliability increased from 10 to 50%.
  • 2.It take 20 trial for epsilon to decay from 1 to 0. the epsilon function was a linear decay of 0.05 it would take 20 trials for the epsilon to decay from 1 to 0. In the parameter value graph, the straight line from 1 to 0 where initial value starts at 0.95 where first computation starts before executing the trail.
  • 3.As no of trial increase the bad action was reduced.yes the average award increases.
  • 4.It remains the same but the rate of reliability has increased but havent exceed the threshold value for being in better rating same with safety though the overall bad acton reduced but rating remains the same.

Improve the Q-Learning Driving Agent

The third step to creating an optimized Q-Learning agent is to perform the optimization! Now that the Q-Learning algorithm is implemented and the driving agent is successfully learning, it's necessary to tune settings and adjust learning paramaters so the driving agent learns both safety and efficiency. Typically this step will require a lot of trial and error, as some settings will invariably make the learning worse. One thing to keep in mind is the act of learning itself and the time that this takes: In theory, we could allow the agent to learn for an incredibly long amount of time; however, another goal of Q-Learning is to transition from experimenting with unlearned behavior to acting on learned behavior. For example, always allowing the agent to perform a random action during training (if $\epsilon = 1$ and never decays) will certainly make it learn, but never let it act. When improving on your Q-Learning implementation, consider the implications it creates and whether it is logistically sensible to make a particular adjustment.

Improved Q-Learning Simulation Results

To obtain results from the initial Q-Learning implementation, you will need to adjust the following flags and setup:

  • 'enforce_deadline' - Set this to True to force the driving agent to capture whether it reaches the destination in time.
  • 'update_delay' - Set this to a small value (such as 0.01) to reduce the time between steps in each trial.
  • 'log_metrics' - Set this to True to log the simluation results as a .csv file and the Q-table as a .txt file in /logs/.
  • 'learning' - Set this to 'True' to tell the driving agent to use your Q-Learning implementation.
  • 'optimized' - Set this to 'True' to tell the driving agent you are performing an optimized version of the Q-Learning implementation.

Additional flags that can be adjusted as part of optimizing the Q-Learning agent:

  • 'n_test' - Set this to some positive number (previously 10) to perform that many testing trials.
  • 'alpha' - Set this to a real number between 0 - 1 to adjust the learning rate of the Q-Learning algorithm.
  • 'epsilon' - Set this to a real number between 0 - 1 to adjust the starting exploration factor of the Q-Learning algorithm.
  • 'tolerance' - set this to some small value larger than 0 (default was 0.05) to set the epsilon threshold for testing.

Furthermore, use a decaying function of your choice for $\epsilon$ (the exploration factor). Note that whichever function you use, it must decay to 'tolerance' at a reasonable rate. The Q-Learning agent will not begin testing until this occurs. Some example decaying functions (for $t$, the number of trials):

$$ \epsilon = a^t, \textrm{for } 0 < a < 1 \hspace{50px}\epsilon = \frac{1}{t^2}\hspace{50px}\epsilon = e^{-at}, \textrm{for } 0 < a < 1 \hspace{50px} \epsilon = \cos(at), \textrm{for } 0 < a < 1$$

You may also use a decaying function for $\alpha$ (the learning rate) if you so choose, however this is typically less common. If you do so, be sure that it adheres to the inequality $0 \leq \alpha \leq 1$.

If you have difficulty getting your implementation to work, try setting the 'verbose' flag to True to help debug. Flags that have been set here should be returned to their default setting when debugging. It is important that you understand what each flag does and how it affects the simulation!

Once you have successfully completed the improved Q-Learning simulation, run the code cell below to visualize the results. Note that log files are overwritten when identical simulations are run, so be careful with what log file is being loaded!

In [45]:
%run smartcab/agent.py
/-------------------------
| Training trial 1
\-------------------------

Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.27)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.70)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.85)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.40)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.96)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.96)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.96)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.88)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded -0.15)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded -0.15)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.55)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.77)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 1.17)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.80)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.90)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.53)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 2
\-------------------------

Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.05)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.50)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.26)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.55)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.61)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.84)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 1.82)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.18)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.88)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.53)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 0.82)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.89)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.89)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.96)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.41)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.20)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.20)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.74)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.38)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.43)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.51)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.49)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.81)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.38)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.86)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.91)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 3
\-------------------------

Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.43)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.77)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.03)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.69)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.80)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.15)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.85)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.48)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.27)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.44)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.91)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.03)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.35)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.46)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.33)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.16)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 0.99)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.19)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 4
\-------------------------

Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.21)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.90)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.13)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.53)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.17)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.36)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.62)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.43)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.81)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.03)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.51)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 1.32)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.29)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.27)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded -0.40)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.07)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.50)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.05)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.11)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 5
\-------------------------

Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.07)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.01)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.81)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.46)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.86)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.77)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.57)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.64)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.63)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.43)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.48)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.55)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.29)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded 1.04)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded -0.17)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.41)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.41)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.29)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.20)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.34)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove left instead of right. (rewarded -0.50)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove left instead of right. (rewarded 0.63)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'right')
Agent drove right instead of left. (rewarded -0.37)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 6
\-------------------------

Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent drove forward instead of left. (rewarded 0.93)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.92)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent followed the waypoint left. (rewarded 2.72)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.08)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.44)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.88)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.44)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.14)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.31)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.35)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.06)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.77)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.77)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded -0.04)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.56)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.93)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.93)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove right instead of left. (rewarded 1.08)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.74)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.99)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.44)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.87)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.20)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.20)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.20)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.20)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.02)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 7
\-------------------------

Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.58)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.62)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.06)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.06)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.15)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.15)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.21)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 0.52)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 1.20)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 1.20)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 1.20)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 1.20)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.01)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.01)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.01)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.01)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.90)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.79)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.79)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.79)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.79)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.55)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'right')
Agent drove right instead of left. (rewarded 0.28)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
57% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
57% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
57% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
57% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.48)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.78)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.38)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.55)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.03)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.30)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.04)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.53)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.29)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.13)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.47)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.60)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.67)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.90)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 8
\-------------------------

Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.06)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.22)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.48)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.81)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.73)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.42)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.19)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.55)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.90)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.64)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.60)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.69)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.89)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.38)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 0.94)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.96)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.90)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.16)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 9
\-------------------------

Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.69)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'right')
Agent drove right instead of left. (rewarded 0.49)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 1.96)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.04)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.12)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.66)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.84)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.02)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.25)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.56)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.56)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.56)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
48% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
48% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
48% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
48% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.25)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.12)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.38)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.78)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.78)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.78)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.99)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.92)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.14)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent drove right instead of left. (rewarded 0.55)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.43)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.21)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 10
\-------------------------

Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.00)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.45)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.95)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.50)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.02)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.36)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.24)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 1.38)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.09)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.79)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.51)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.57)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.49)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 11
\-------------------------

Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.82)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.31)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.43)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.28)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.79)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.40)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.33)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.55)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.77)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 1.62)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.79)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.38)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.50)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.03)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.38)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.77)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.52)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.02)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 12
\-------------------------

Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.87)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.84)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.19)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.49)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.88)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.66)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.30)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.30)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.30)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.30)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.30)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.78)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.17)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.99)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.66)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.33)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.76)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.46)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.29)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.61)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.61)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.61)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.26)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.47)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 13
\-------------------------

Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent drove forward instead of left. (rewarded 0.70)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.03)
92% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.03)
92% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.03)
92% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.03)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.54)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.88)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.54)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.97)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.97)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.97)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.97)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.47)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.21)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.69)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 1.71)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.69)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.56)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove forward instead of right. (rewarded 0.29)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.37)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.47)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.79)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded -0.30)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.71)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.38)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.26)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 14
\-------------------------

Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.36)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.74)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.21)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.69)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.21)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.02)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 2.72)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.84)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 0.33)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.53)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.11)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.03)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.03)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.03)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.03)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.03)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.03)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.03)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.03)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.50)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.11)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.11)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.11)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.11)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.26)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.06)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.02)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.23)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.86)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.07)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.07)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.07)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 15
\-------------------------

Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.37)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.70)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.44)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.02)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.88)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.01)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.17)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.36)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.35)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.27)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.96)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.52)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.00)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.43)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.68)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.55)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.80)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 16
\-------------------------

Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.88)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.15)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.41)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.96)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.52)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded -0.12)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.53)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.94)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.46)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.38)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.40)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.39)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.08)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.08)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.08)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.67)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.10)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove right instead of left. (rewarded -0.67)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 1.13)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 17
\-------------------------

Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.00)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.91)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.70)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.70)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.70)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.34)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.07)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 1.48)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.74)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.74)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.61)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove forward instead of right. (rewarded 0.38)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.96)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.60)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.33)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.35)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.35)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent drove right instead of left. (rewarded 0.10)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.46)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.46)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.46)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.46)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.17)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.55)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 18
\-------------------------

Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.97)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.00)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.89)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.89)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.28)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.55)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.20)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.27)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.49)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.29)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.43)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.41)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.41)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 0.10)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 0.71)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.69)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.85)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.58)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.87)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.86)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.18)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 19
\-------------------------

Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.25)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.85)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.14)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.79)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.11)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.45)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.45)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.45)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.03)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.27)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.66)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.52)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.62)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.54)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.48)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.19)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.39)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded -0.08)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded -0.08)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.43)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 20
\-------------------------

Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.56)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.25)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent drove forward instead of right. (rewarded 0.50)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.96)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.17)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.28)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.70)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.13)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.94)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.54)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.45)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.55)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.55)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.16)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.16)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.16)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.36)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.47)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.15)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.38)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 21
\-------------------------

Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.73)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.33)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.33)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.33)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.40)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.53)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.61)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.75)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.17)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.09)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.09)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.09)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.43)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.91)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.91)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.36)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.87)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.20)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.26)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.26)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.26)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.26)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.26)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.26)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.36)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 22
\-------------------------

Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.66)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.62)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.36)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.99)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.99)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.99)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.99)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.36)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.86)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.65)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.07)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.05)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.34)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.70)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 0.68)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.09)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.02)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.47)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.47)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.47)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.47)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent drove right instead of left. (rewarded -0.14)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.63)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 23
\-------------------------

Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.99)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.50)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.37)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent drove left instead of right. (rewarded 1.03)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.12)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.12)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.12)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.12)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.12)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.32)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.67)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.67)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.67)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.67)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.67)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.20)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.35)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.69)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.69)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.69)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.69)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.85)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.59)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.59)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.59)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.59)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.59)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.59)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.59)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.59)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.05)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.68)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.56)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 24
\-------------------------

Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.31)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.04)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.04)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.04)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.04)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.04)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.32)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.44)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.83)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.76)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.64)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.41)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.13)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.65)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.09)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent drove right instead of left. (rewarded 0.58)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 1.55)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent followed the waypoint forward. (rewarded 0.71)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.70)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.51)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.73)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.73)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.73)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.73)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.73)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 1.85)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 25
\-------------------------

Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.29)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.03)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.11)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 0.96)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.33)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.85)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.52)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.23)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.23)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.23)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.23)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.62)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.21)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.21)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.21)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.21)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.21)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.07)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.80)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.16)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.42)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.13)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.95)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.44)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.04)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.87)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.27)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.47)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.77)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -9.63)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.62)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.93)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.02)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.25)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.79)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 26
\-------------------------

Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.31)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.39)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove forward instead of right. (rewarded 1.75)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.49)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.88)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.84)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.83)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.12)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.21)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.21)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.21)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.82)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.77)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.77)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.77)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.00)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.41)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.05)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.62)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.62)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.83)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 27
\-------------------------

Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.52)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.11)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.47)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.55)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.55)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.55)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.66)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.73)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.73)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 1.51)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 1.51)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 1.51)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.66)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.24)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.36)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.17)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.11)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'right')
Agent drove right instead of left. (rewarded 0.90)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.96)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 28
\-------------------------

Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.53)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.37)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.03)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.92)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.34)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.30)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.71)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.46)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.03)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.46)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.46)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.46)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.46)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.46)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.26)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.26)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.26)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.15)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.84)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.36)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.39)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.91)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 29
\-------------------------

Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.52)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.44)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.10)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.25)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove left instead of right. (rewarded 1.08)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.06)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.06)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.06)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.84)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.54)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.62)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.05)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.02)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 0.86)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.02)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.18)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.43)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 0.91)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.16)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.41)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.22)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.64)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 30
\-------------------------

Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.36)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.28)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.51)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.13)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.42)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.86)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.92)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.41)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.48)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.96)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.18)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.23)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.31)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.00)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 0.63)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 1.07)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 31
\-------------------------

Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.14)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.97)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.29)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.11)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.11)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.03)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.81)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.11)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.13)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.42)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.42)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.42)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.64)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.40)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.56)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded -0.15)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.50)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.01)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.91)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.08)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded -0.55)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.19)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 32
\-------------------------

Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.41)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.82)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.92)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.10)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.04)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.50)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.09)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.65)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.22)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.63)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.65)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.65)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.35)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.89)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.89)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.89)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.89)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.54)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 1.17)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.32)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.32)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.40)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.73)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded -0.27)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded -0.27)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded -0.27)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.07)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.71)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.71)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.71)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.55)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.71)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.62)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 0.66)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 33
\-------------------------

Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.57)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.68)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.96)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.96)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.96)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.31)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.53)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.74)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.52)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.28)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.51)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.28)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.11)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.29)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.44)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.47)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.05)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.48)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.63)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.55)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.37)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.37)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.37)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.37)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.37)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -10.95)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 34
\-------------------------

Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -9.54)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.76)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.91)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.86)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.33)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.07)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.99)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.09)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.51)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.71)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.71)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.71)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.71)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.71)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded -0.16)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.73)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.44)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.83)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.56)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.80)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.84)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.07)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.39)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.60)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.95)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 35
\-------------------------

Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent drove forward instead of left. (rewarded 1.03)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.78)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.50)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.87)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.26)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.01)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded -0.10)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.60)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.87)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.91)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.83)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.80)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.52)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.89)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.30)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.92)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.34)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.14)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.06)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 36
\-------------------------

Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.61)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.02)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.87)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.05)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.43)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.91)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.30)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.13)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.08)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.99)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.31)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.06)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.22)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.46)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.37)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.77)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.05)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.37)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded -0.41)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 37
\-------------------------

Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.53)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.63)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded -0.04)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.03)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.94)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 1.37)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.38)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.88)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.36)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.47)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.72)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.18)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.32)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.32)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded -0.13)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove right instead of left. (rewarded -0.61)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.59)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.28)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 38
\-------------------------

Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.42)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.65)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.65)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.17)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.77)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.10)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.02)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.04)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.04)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.04)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.33)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.01)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.01)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.01)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.27)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.27)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.27)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.49)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.49)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.49)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.49)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.85)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.44)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.18)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.18)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.18)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 0.46)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.11)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.11)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.11)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.11)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.11)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.10)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.54)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.59)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.78)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.00)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 39
\-------------------------

Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.48)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.73)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.78)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.29)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.51)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.49)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.69)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.59)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.43)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent followed the waypoint right. (rewarded 2.39)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.75)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.37)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.85)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.85)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.72)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.90)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.70)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 1.24)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.14)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.59)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.13)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.00)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 1.29)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.65)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.77)
17% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 40
\-------------------------

Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.43)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.05)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.71)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.02)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.57)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.81)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove left instead of forward. (rewarded 0.42)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.82)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.77)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 41
\-------------------------

Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.60)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.60)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.60)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.60)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.60)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.82)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.98)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.98)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.98)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.98)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.98)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.98)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.98)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.39)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.46)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.42)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.30)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.24)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.32)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.34)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.51)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.51)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.30)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.34)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.75)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.57)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.57)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.57)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.57)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.87)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.21)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 42
\-------------------------

Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.10)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 0.12)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent drove right instead of left. (rewarded 0.74)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.33)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.78)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.46)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.62)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 1.57)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.11)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 43
\-------------------------

Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.66)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded 1.90)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.32)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.26)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.76)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.01)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.08)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.29)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.77)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.77)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.32)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.26)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.16)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.61)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded -0.24)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded -0.24)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded -0.24)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded -0.18)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.80)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.28)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.86)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 44
\-------------------------

Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.14)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.49)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.50)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.36)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.69)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.27)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.49)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.92)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.57)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.16)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.64)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.77)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.63)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.01)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.33)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.33)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.33)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.41)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.07)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.07)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.40)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.97)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.41)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 45
\-------------------------

Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.81)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.85)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.52)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.68)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.88)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 1.56)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent drove right instead of forward. (rewarded 0.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent drove right instead of forward. (rewarded 0.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent drove right instead of forward. (rewarded 0.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent drove right instead of forward. (rewarded 0.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent drove right instead of forward. (rewarded 0.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent drove right instead of forward. (rewarded 0.02)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.02)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.15)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.86)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded -0.02)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded -0.02)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded -0.02)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded -0.02)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded -0.02)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded -0.02)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.85)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.40)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.19)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.64)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.94)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.71)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.22)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.22)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.22)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.55)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.64)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove forward instead of right. (rewarded 0.30)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.14)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.05)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.05)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.05)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.05)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.97)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.97)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.25)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.25)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.25)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.25)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.25)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.25)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.25)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.25)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.14)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
14% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
14% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
14% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.60)
9% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.60)
9% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.60)
9% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.60)
9% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.60)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.41)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.35)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 0.70)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 46
\-------------------------

Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.83)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.38)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.19)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.47)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.65)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.06)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent followed the waypoint forward. (rewarded 1.76)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent followed the waypoint forward. (rewarded 1.76)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent followed the waypoint forward. (rewarded 1.76)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent followed the waypoint forward. (rewarded 1.76)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.43)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.18)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.39)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.39)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.39)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.39)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'left')
Agent drove forward instead of left. (rewarded 1.10)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.13)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.59)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.29)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.49)
49% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.49)
49% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.49)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.65)
46% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.65)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.48)
43% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.48)
43% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.48)
43% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.48)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.08)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.70)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.04)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.66)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.61)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.53)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.25)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.09)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.05)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.11)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.52)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.78)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.34)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.97)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 47
\-------------------------

Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.22)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.54)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.19)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 1.42)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.82)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.03)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.81)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.73)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.48)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.41)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.49)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.16)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent followed the waypoint right. (rewarded 2.10)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent followed the waypoint right. (rewarded 2.10)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.88)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.88)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.88)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.88)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.05)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.82)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 48
\-------------------------

Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.27)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.09)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.98)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.90)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.46)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.12)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.73)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.73)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 0.91)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.79)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.65)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.67)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.59)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.42)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.19)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.19)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.19)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.19)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.19)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.19)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.93)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded -0.36)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.22)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.67)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded -0.05)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.34)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 49
\-------------------------

Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.55)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.47)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.20)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.20)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.91)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.88)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.35)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.02)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.85)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.33)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.30)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.30)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.30)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.30)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.84)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.29)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.29)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 0.75)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.17)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.48)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.28)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.70)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 0.35)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 50
\-------------------------

Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.50)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.89)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.71)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent drove forward instead of right. (rewarded 1.85)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.03)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.62)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent drove right instead of forward. (rewarded 1.72)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.61)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.40)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.26)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.50)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.30)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.48)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.49)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.49)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.49)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.96)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.14)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.45)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 51
\-------------------------

Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.88)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.39)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.45)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.84)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.84)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.84)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.84)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.84)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.84)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.84)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.49)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.04)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent drove right instead of forward. (rewarded 1.19)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.46)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.64)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.49)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.49)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.49)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.21)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.13)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.40)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.38)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.97)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.75)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.80)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'forward')
Agent properly idled at a red light. (rewarded 0.89)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.50)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 52
\-------------------------

Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.73)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.70)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.91)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.43)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.33)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.78)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.88)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded 1.69)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.89)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.02)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.83)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.13)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.90)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.08)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 53
\-------------------------

Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.02)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 2.91)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.04)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.06)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.39)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.05)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.63)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.42)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.22)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.04)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.04)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.36)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.46)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.50)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.95)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.17)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.17)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.03)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.01)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 0.58)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.32)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.70)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.23)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.29)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.98)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.68)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.38)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.37)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 54
\-------------------------

Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.09)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 1.17)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.93)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.04)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.91)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.03)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.21)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.62)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.62)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.62)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.58)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.94)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.93)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.54)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 1.58)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.10)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.30)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.51)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.60)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.81)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.18)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.10)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.34)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded -0.10)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.65)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.90)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.49)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.49)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.01)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.31)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 55
\-------------------------

Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.08)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.73)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.86)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.32)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.76)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.16)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.36)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.17)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.26)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.26)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.54)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.94)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 0.80)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 0.80)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 0.80)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.83)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.42)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.28)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.61)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.73)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.05)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.52)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.52)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.44)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.44)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.20)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.30)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 56
\-------------------------

Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.19)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.25)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.37)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.11)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 2.40)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.51)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.38)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.29)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.13)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.73)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 0.92)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.20)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.28)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 57
\-------------------------

Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.19)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.11)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'right')
Agent drove right instead of left. (rewarded 1.11)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 1.35)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.29)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.54)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.52)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.22)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 1.25)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.01)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.86)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.25)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.61)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.12)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.51)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.47)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 58
\-------------------------

Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.03)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.92)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.10)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.05)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.52)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.03)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.59)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.17)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.81)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.54)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.29)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.44)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.99)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 1.23)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 1.23)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.44)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.16)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.16)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.35)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.35)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.51)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.27)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.91)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove right instead of forward. (rewarded -0.56)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 59
\-------------------------

Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.59)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.80)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.76)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.38)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.43)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent drove right instead of forward. (rewarded 0.75)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.80)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.96)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.96)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.96)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.45)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.12)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.27)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.48)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.76)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.05)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.05)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.33)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.19)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 60
\-------------------------

Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 1.73)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.99)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.25)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.28)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 61
\-------------------------

Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.05)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 0.53)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.30)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.06)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.00)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.49)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.12)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.35)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.60)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.38)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.61)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.04)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.38)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.11)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.11)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.11)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.28)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.62)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.09)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.09)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.37)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.46)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.53)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.18)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.64)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.86)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.72)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.90)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 62
\-------------------------

Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.44)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.89)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.33)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.58)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.52)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.52)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.52)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.32)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.32)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.32)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.32)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.32)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent followed the waypoint forward. (rewarded 2.78)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.51)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.51)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.51)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.51)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.51)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.51)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.51)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.35)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.12)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.06)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.30)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.09)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.09)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.09)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 1.53)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.70)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.11)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.65)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.80)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.04)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.04)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.04)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.04)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.04)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.04)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.04)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.43)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.52)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.96)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.55)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded -0.85)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 63
\-------------------------

Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.32)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.62)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.24)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.51)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.75)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.05)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.51)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.59)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.32)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.65)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.20)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded -0.03)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.35)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.50)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.50)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.50)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.07)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.38)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 0.90)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.49)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 64
\-------------------------

Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.20)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.71)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.15)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.59)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.10)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.95)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.95)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.95)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.95)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.95)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.46)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.63)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.55)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.13)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.02)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.31)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.17)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.17)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.31)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent followed the waypoint right. (rewarded 1.55)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded -0.00)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.50)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.86)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.05)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 1.14)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.42)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.53)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 0.82)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 65
\-------------------------

Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.43)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.83)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.53)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.38)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.52)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.09)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.09)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.09)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 2.20)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.06)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.98)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.52)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.12)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove forward instead of left. (rewarded 0.95)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.44)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.97)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.20)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.49)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.59)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.07)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.26)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.96)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.70)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.37)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.79)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove left instead of right. (rewarded 0.53)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.00)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.17)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 66
\-------------------------

Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.88)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.11)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.27)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.30)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.07)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.63)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.38)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.95)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.46)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove right instead of forward. (rewarded 0.60)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.04)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent followed the waypoint right. (rewarded 2.74)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.29)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.54)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.55)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.53)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.92)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.19)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.34)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.30)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.87)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.97)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded -0.72)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 67
\-------------------------

Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.86)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.74)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.20)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.20)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.80)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.75)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.86)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.13)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.82)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.55)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.43)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.57)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.59)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.59)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.10)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.82)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.20)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.36)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.29)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.92)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.17)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.20)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 68
\-------------------------

Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.13)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.41)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.94)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.64)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.54)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.07)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.88)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.93)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded -0.10)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded -0.10)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded -0.10)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded -0.10)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.18)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.41)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.92)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.88)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.32)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 69
\-------------------------

Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.79)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.80)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.10)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.74)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.47)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.48)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.03)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.97)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.30)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.49)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.98)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.08)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.74)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.69)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.93)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.44)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.43)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.42)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 1.50)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.72)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded -0.36)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.15)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.50)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.27)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.15)
6% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.15)
6% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.15)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.48)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 70
\-------------------------

Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.81)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 1.52)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.90)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.93)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.93)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.93)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.93)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.93)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.93)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.53)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.23)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.69)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.43)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 1.43)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.59)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.33)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.81)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.59)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.64)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove right instead of forward. (rewarded -0.59)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.40)
5% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 71
\-------------------------

Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.82)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.31)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.49)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.50)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.58)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.58)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.40)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.40)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.87)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.86)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.64)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.93)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.21)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.69)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.38)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.94)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.22)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded -0.44)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.90)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded -0.40)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded -0.40)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove forward instead of right. (rewarded 0.91)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.70)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 72
\-------------------------

Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.95)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.91)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.77)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.13)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.56)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.33)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.33)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.33)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.33)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.68)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.98)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.98)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.98)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.98)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.98)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.98)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.98)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.96)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.39)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.12)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.12)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.92)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.90)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.53)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.38)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.70)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.80)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.00)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 1.26)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.33)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.04)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.04)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.04)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.04)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.04)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.04)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.40)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.43)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.41)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.41)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.41)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.41)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.41)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.41)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.41)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.30)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 73
\-------------------------

Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.24)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.06)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.87)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.78)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'right', None)
Agent drove right instead of left. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'right', None)
Agent drove right instead of left. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'right', None)
Agent drove right instead of left. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'right', None)
Agent drove right instead of left. (rewarded 0.95)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.13)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.89)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.81)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.68)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.90)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.98)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.24)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.24)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.24)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.24)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.24)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.24)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.24)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.68)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.57)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 74
\-------------------------

Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.97)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.41)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.29)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.49)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.49)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.11)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.46)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.31)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.75)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.67)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.32)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.42)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.07)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.53)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.53)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.98)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.71)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.23)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.67)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.25)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.25)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.25)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.36)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.98)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 75
\-------------------------

Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.84)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.12)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.31)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.52)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.37)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.99)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 1.23)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.54)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.18)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.48)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.18)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.04)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.82)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.69)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.02)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.95)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 76
\-------------------------

Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.85)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.29)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.19)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.59)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.88)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.18)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.18)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.18)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.18)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.18)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove left instead of forward. (rewarded 1.20)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.77)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.63)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.13)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.43)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.62)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 1.29)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.71)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.48)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove forward instead of right. (rewarded 1.53)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.10)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.10)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.31)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.61)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.26)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.35)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded -0.45)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 0.76)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.07)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded -0.05)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.23)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 77
\-------------------------

Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.87)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.59)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.23)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.23)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.23)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.23)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.23)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded -0.02)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.99)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.43)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.62)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.61)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.16)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.56)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.76)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.33)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.33)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.33)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.33)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 1.13)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 1.13)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 1.13)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent drove right instead of left. (rewarded -0.07)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.82)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.55)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded -0.21)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.44)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 78
\-------------------------

Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.45)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.47)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.40)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.69)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.17)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.70)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.85)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.18)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.03)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.52)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent drove forward instead of left. (rewarded 0.69)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.50)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.50)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.04)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.06)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.10)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.51)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.57)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.53)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.63)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.78)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent drove right instead of forward. (rewarded -0.46)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.22)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded -0.30)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 79
\-------------------------

Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.36)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
92% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.59)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.77)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.86)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.80)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.76)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent drove right instead of left. (rewarded 1.65)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.61)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.27)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.10)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.57)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.12)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.00)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.06)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.81)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'right')
Agent followed the waypoint right. (rewarded 1.08)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.36)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.67)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.17)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.78)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.04)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 80
\-------------------------

Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.09)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.36)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.31)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.96)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.04)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.94)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.94)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.01)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.21)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.29)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.67)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.81)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.42)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.98)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.57)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.76)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.89)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.55)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.90)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.26)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.10)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.95)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.14)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.66)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 81
\-------------------------

Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.72)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.64)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.43)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.27)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.33)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.71)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.06)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.60)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.37)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.65)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.08)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.17)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 82
\-------------------------

Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.15)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.43)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.78)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.02)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.32)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.25)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.84)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.84)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.84)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.84)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.84)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.84)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.08)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.38)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.66)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.80)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.91)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.94)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.96)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.47)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 1.28)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.57)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.51)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.83)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.17)
23% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 83
\-------------------------

Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.40)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', None)
Agent drove right instead of left. (rewarded 0.85)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.11)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.11)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.11)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.11)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.02)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.64)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.21)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.67)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.24)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.28)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.56)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.33)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.27)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.47)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.59)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.59)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.59)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.59)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.59)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.67)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.36)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.61)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.42)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.27)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.09)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.44)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.44)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.69)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent followed the waypoint right. (rewarded 1.12)
17% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 84
\-------------------------

Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove forward instead of right. (rewarded 1.42)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.33)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.65)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.34)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.34)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.82)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.48)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.86)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.43)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.56)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.60)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.09)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 1.45)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.29)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.04)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.62)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.55)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded -0.06)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.28)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.59)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.88)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.14)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 85
\-------------------------

Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', 'right')
Agent drove right instead of left. (rewarded 0.12)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.90)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.40)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.61)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.82)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.47)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.68)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.16)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.16)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.16)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.16)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.08)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.87)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.86)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.00)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 86
\-------------------------

Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.69)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.74)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.83)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.63)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.63)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.63)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.96)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.09)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.16)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.18)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.26)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.26)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.26)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.26)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.26)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.26)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.78)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.83)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.83)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.83)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.87)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.80)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.67)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.32)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.55)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.40)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.38)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.24)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.37)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.46)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.99)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.57)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.34)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.74)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 87
\-------------------------

Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.68)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.14)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.78)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.21)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 1.57)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.09)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.09)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.09)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.09)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.54)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.75)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.44)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.31)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.40)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.57)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.03)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.23)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.26)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.51)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 88
\-------------------------

Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.67)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
92% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
92% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.20)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.94)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.55)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.21)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.48)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.06)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.30)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.23)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.39)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.33)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.91)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.91)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.91)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.83)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.50)
36% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 89
\-------------------------

Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.54)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.34)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.95)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.90)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.90)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.25)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.04)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 1.73)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.91)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.75)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.06)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.80)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.55)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.76)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 90
\-------------------------

Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 1.76)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove forward instead of right. (rewarded 0.94)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 1.78)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.83)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.02)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.26)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded -0.02)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.29)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.45)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.44)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.68)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.68)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.68)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.68)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.68)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.68)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.68)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.23)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.08)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove forward instead of right. (rewarded 0.47)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove forward instead of right. (rewarded 0.47)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove forward instead of right. (rewarded 0.47)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove forward instead of right. (rewarded 0.47)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.61)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.35)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.83)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 91
\-------------------------

Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.07)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.05)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.84)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.77)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.28)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.59)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.72)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.72)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.71)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 92
\-------------------------

Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.45)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.09)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.09)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.87)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.30)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove forward instead of right. (rewarded 1.50)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent followed the waypoint right. (rewarded 2.59)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.91)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.53)
63% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 93
\-------------------------

Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.74)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.65)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.03)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.81)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.26)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.37)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.19)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.85)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.37)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.15)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.54)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.54)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.34)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.80)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.24)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.19)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.02)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.63)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.42)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.17)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.65)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.09)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 1.15)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.17)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.96)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.77)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 94
\-------------------------

Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.85)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.98)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.83)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.45)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.57)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.06)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.81)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.63)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.95)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.96)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.53)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.48)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.58)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.39)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.58)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.12)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.64)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.52)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent followed the waypoint left. (rewarded 1.42)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.49)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.49)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.32)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.45)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 95
\-------------------------

Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.67)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.93)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.36)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.52)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent drove forward instead of right. (rewarded -0.07)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.83)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.83)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.83)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.83)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.41)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.07)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.29)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.78)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.54)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.13)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 0.33)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.91)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.91)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.91)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.91)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.91)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.53)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 1.00)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.76)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded -0.17)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 96
\-------------------------

Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.21)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.49)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.96)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.35)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.35)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.74)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.53)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.07)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.26)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.04)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.34)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.05)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.47)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded -0.67)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.11)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.41)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 97
\-------------------------

Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.80)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.54)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.77)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.73)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.55)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.32)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 98
\-------------------------

Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent drove forward instead of right. (rewarded 1.09)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.04)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.44)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.55)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.19)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.70)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.15)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded 1.20)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.20)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.53)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.46)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.40)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.90)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.07)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.07)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.07)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.50)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.23)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.08)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.48)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.48)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.48)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.79)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.17)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 99
\-------------------------

Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.29)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.45)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.15)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.91)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.88)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.31)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.00)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.90)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.60)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.61)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.93)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.56)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.81)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.08)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.61)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.64)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.68)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.48)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.95)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.04)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.77)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.26)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded -0.46)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.53)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 100
\-------------------------

Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.82)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove forward instead of left. (rewarded 1.45)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.62)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.18)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.48)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.91)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.45)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.25)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.59)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.40)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.69)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.19)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.38)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.48)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.48)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.00)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.70)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.37)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.70)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded -0.22)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.21)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.72)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.09)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.91)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 101
\-------------------------

Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.21)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.78)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.05)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.51)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 1.85)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.51)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.25)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.58)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.55)
71% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.55)
71% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.55)
71% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.55)
71% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.55)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.22)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.22)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.10)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.73)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 1.17)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.41)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.62)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.88)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent drove right instead of forward. (rewarded 1.12)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.38)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.74)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.16)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.66)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.42)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.44)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.73)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.53)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.09)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.95)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 0.33)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.43)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.94)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.16)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.04)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 102
\-------------------------

Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.94)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.54)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.54)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.54)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.54)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.54)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.54)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.54)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.54)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.84)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.54)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'right')
Agent properly idled at a red light. (rewarded 0.71)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.91)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.35)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 103
\-------------------------

Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.35)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.48)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.48)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.53)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.53)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.53)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.53)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.61)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.26)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.30)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.13)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent drove right instead of left. (rewarded 1.86)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.59)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.60)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.20)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.10)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.70)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'right', 'right')
Agent drove right instead of left. (rewarded 0.07)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 1.52)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.36)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.36)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded -0.12)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.69)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.91)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.88)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.04)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 104
\-------------------------

Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.26)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.85)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.44)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.95)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.39)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.53)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.33)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.80)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.83)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 105
\-------------------------

Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.34)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.76)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.09)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.43)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.35)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.70)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.67)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.99)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.93)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.32)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.32)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.32)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.32)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.32)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.32)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.32)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.35)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.56)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.56)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.56)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.56)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.56)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.56)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.56)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 0.70)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.23)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.30)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.79)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.47)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.77)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.17)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded -0.00)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.14)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded -0.04)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.32)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.01)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.36)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.69)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.69)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent followed the waypoint right. (rewarded 0.18)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 106
\-------------------------

Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.80)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 1.50)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.81)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.79)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.97)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.48)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.27)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.84)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.36)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.62)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.13)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'right')
Agent drove right instead of left. (rewarded 0.40)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.14)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.27)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.67)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 1.82)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.78)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.33)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.95)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.97)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.02)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.42)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.71)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.26)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.48)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.15)
17% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 107
\-------------------------

Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.84)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.28)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.50)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.50)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.50)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.56)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.31)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.31)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.31)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.61)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.61)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.02)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.21)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.20)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.22)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.37)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.75)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.48)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.91)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.91)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.26)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 0.12)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 0.88)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 108
\-------------------------

Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 1.38)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.12)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.52)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 0.45)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.12)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.78)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.48)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 109
\-------------------------

Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.92)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.03)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.21)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.33)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.26)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.43)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.30)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.30)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.30)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.30)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.30)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.30)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.30)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.31)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.25)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.08)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.67)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.67)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.48)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded -0.41)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.60)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 110
\-------------------------

Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 1.78)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 0.07)
92% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 0.07)
92% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 0.07)
92% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 0.07)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.17)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.40)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.42)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.03)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.03)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.62)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent followed the waypoint forward. (rewarded 1.13)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.01)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.68)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.40)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.53)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.31)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.97)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.83)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.31)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.44)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 0.58)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 0.58)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 0.58)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.93)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.75)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.84)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 111
\-------------------------

Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.89)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.05)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.65)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.85)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.85)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.85)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.85)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.85)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.03)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.52)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.78)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.17)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.86)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.54)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.28)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.50)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.80)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.80)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.41)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded -0.13)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.36)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.11)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.04)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 112
\-------------------------

Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.34)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.34)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.57)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.57)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.07)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.72)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 1.59)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded -0.07)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.39)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.85)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 113
\-------------------------

Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.91)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -9.68)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.81)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.81)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.91)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.19)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.94)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.19)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.11)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.11)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.11)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.45)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.83)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.16)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.26)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.04)
28% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.04)
28% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.04)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.28)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.32)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.99)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.99)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.99)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.99)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.99)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.99)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 0.76)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.06)
8% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 114
\-------------------------

Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.75)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.96)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.53)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.27)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.27)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.27)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.28)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 1.59)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 0.35)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.34)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.22)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.85)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.24)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.92)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.23)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.74)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 1.03)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.02)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.41)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.41)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.41)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 0.65)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.13)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.80)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.44)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded -0.73)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 115
\-------------------------

Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.09)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.59)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.40)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.09)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.20)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.88)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.71)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.34)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.63)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.32)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded -0.06)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.50)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 116
\-------------------------

Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.80)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.68)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.44)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.04)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.04)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.04)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.04)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.04)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.04)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.04)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.04)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.32)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.64)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.69)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.33)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.57)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.25)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.42)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded -0.23)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.60)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.65)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.52)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.52)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.52)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.52)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.52)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.52)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 1.32)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 0.70)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.48)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded -0.25)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.28)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.51)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.51)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.51)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.51)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.51)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.51)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.51)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.51)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.23)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.98)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 117
\-------------------------

Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.00)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.47)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.14)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.71)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.92)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.89)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.62)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.99)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.87)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.51)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 118
\-------------------------

Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.05)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.01)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.29)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.29)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.29)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.67)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.60)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.53)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.79)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.79)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.56)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.31)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.52)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 1.26)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.27)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.59)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove left instead of forward. (rewarded 1.08)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.29)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.29)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.29)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.29)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.29)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.87)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.32)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.77)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.66)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.84)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.18)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.84)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 119
\-------------------------

Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent followed the waypoint right. (rewarded 1.72)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.31)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.64)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.93)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.89)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.02)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.68)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.89)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.84)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.48)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.45)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.94)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.30)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.31)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.34)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.01)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.01)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.01)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.02)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 120
\-------------------------

Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.76)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.19)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.15)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.31)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.31)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.31)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.31)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.31)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.31)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.48)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.50)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.13)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.49)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.49)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 0.95)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.45)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.02)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 121
\-------------------------

Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.58)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 1.42)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 1.34)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.87)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.87)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.18)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.27)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.27)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.27)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.59)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.59)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.59)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.55)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.79)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.89)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.72)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.62)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.42)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 122
\-------------------------

Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove forward instead of right. (rewarded 1.71)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.82)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.44)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.58)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded -0.08)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.64)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.03)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.28)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.18)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.56)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.32)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.14)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.65)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.13)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.13)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.13)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.13)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.33)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.35)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.15)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.05)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.35)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 123
\-------------------------

Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.43)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.94)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.86)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.34)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.34)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.10)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.66)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'left')
Agent followed the waypoint left. (rewarded 1.60)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.23)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.36)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded -0.14)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.94)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.62)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.58)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.43)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.78)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.75)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 124
\-------------------------

Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.77)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.52)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.02)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.61)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.34)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.72)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.72)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded -0.13)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.23)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.30)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.59)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 125
\-------------------------

Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.43)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.03)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.49)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.49)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.49)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.49)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.49)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.60)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.54)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.86)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.98)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.98)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.95)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.11)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.17)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 126
\-------------------------

Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.77)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.38)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.61)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.02)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.90)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.29)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.29)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.29)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.29)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.29)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.05)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.04)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent followed the waypoint right. (rewarded 1.66)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
63% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.98)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.98)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.98)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.98)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.79)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.13)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.74)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.98)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.75)
37% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 127
\-------------------------

Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.99)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.84)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.46)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.21)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.55)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.83)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.96)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.45)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.32)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 128
\-------------------------

Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.24)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.08)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.38)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.67)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.67)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.67)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.67)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.67)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.67)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.67)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.67)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.86)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.34)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 0.21)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.98)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.16)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 1.76)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.73)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.13)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.13)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.13)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.64)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.79)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.79)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.79)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.79)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.79)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.79)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded -0.25)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.99)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.18)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.10)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.10)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.10)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.10)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.10)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.10)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.10)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded -0.57)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.45)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.40)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded -0.05)
7% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded -0.05)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.80)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 129
\-------------------------

Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.14)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.00)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.88)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 1.75)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.79)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.74)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.49)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.07)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.99)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.83)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.95)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.43)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.55)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.16)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.25)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded -0.35)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.01)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.05)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.17)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 130
\-------------------------

Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent drove right instead of forward. (rewarded 0.91)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.79)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.34)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.30)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.82)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.23)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.78)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.65)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.68)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.89)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.89)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.89)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.89)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.89)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.89)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.01)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove forward instead of left. (rewarded 1.77)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.59)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.19)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.44)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.44)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.44)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.44)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.44)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.44)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.41)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.80)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.80)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.80)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.80)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.80)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.15)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.49)
23% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 131
\-------------------------

Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.10)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.14)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.72)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.95)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.61)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded -0.03)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.47)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.08)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.92)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.68)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent drove forward instead of right. (rewarded 1.39)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.28)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.65)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.09)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.54)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.22)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.36)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.35)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.47)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.54)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.22)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.64)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.97)
34% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.97)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.42)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.81)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.62)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.71)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.71)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.60)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.60)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.01)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.52)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.69)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.96)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 132
\-------------------------

Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.36)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.23)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 1.21)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.80)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.81)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.55)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded -0.02)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.78)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.24)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.16)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded -0.10)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded -0.10)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded -0.10)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.45)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.45)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.45)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.45)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.83)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.15)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.15)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.15)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.15)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.15)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.15)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.15)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.15)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.22)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.67)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.67)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.67)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.31)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
33% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
33% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
33% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
33% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
33% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
33% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
33% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.20)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.19)
13% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 133
\-------------------------

Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.44)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.73)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.70)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.19)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 1.56)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.23)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.77)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.12)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.38)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.41)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.03)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 134
\-------------------------

Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.16)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.59)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.59)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.59)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.59)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.59)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.59)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.59)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.59)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.00)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.23)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.60)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.60)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.60)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.47)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.47)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.47)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.47)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.99)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent followed the waypoint forward. (rewarded 0.87)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.12)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.12)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.12)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.12)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.87)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.52)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.51)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.51)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.51)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.51)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.51)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.51)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.51)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.84)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.56)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.21)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.04)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 135
\-------------------------

Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.22)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.72)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.44)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.16)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.20)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.52)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.08)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.08)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.64)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.84)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.55)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded -0.18)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
48% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
48% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
48% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 0.74)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.18)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.26)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.15)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.14)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.62)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.36)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.36)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.43)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.78)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.65)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.46)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 136
\-------------------------

Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.17)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 1.82)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.68)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.83)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.88)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.25)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.67)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.10)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.72)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.69)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.60)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.86)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.26)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.26)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.37)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.56)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.19)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded -0.65)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 137
\-------------------------

Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.56)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.55)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent drove right instead of left. (rewarded 0.57)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.64)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.50)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.36)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.54)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.52)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 1.89)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.19)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.58)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.87)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded -0.19)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.96)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.47)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.84)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 138
\-------------------------

Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.98)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.29)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.74)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.74)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.74)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.74)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.74)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.74)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.74)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.81)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.77)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.80)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.80)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.80)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.80)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.69)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.03)
63% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 139
\-------------------------

Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent followed the waypoint right. (rewarded 2.73)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.08)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.72)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.62)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.62)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.62)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.56)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.51)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.92)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.04)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.64)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.20)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.64)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.93)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.20)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.73)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.07)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.05)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.30)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.30)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.30)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.30)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.30)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.20)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.60)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.26)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.13)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.48)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.28)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 0.87)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.45)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.31)
14% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.31)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.26)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.11)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.31)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.43)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 140
\-------------------------

Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.74)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.28)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.28)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.28)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.40)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.00)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.74)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.66)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 141
\-------------------------

Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.70)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.82)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.98)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.21)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.20)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.47)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.31)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.40)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.56)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 142
\-------------------------

Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.03)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.05)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.18)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.21)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded 0.14)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.59)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.74)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.54)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 143
\-------------------------

Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.59)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.87)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.59)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.30)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.05)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.58)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.73)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.73)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.73)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.17)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.80)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.64)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.55)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.82)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.20)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.11)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.32)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.67)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 144
\-------------------------

Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.45)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.97)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.61)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.50)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.60)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.62)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.80)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.39)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.25)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.78)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.72)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.27)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.87)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.03)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.03)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.03)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.41)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.44)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.07)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove forward instead of right. (rewarded -0.28)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.00)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.07)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.45)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.45)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.51)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.51)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.51)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.51)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.09)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.35)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.99)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.46)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.28)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 145
\-------------------------

Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.07)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.92)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.32)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.80)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.20)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.09)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.17)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.11)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.77)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent drove right instead of left. (rewarded 0.18)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.69)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.66)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.70)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.63)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.09)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.38)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded -0.19)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.40)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.03)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.82)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.22)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.76)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.37)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 0.35)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.27)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.63)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.90)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 146
\-------------------------

Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.53)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.93)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.62)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.39)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 0.10)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.25)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.48)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.63)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.51)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.02)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.06)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.01)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.32)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.75)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.61)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded -0.57)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.42)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.58)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 147
\-------------------------

Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.49)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.77)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.57)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.01)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.38)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.62)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.89)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.07)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.62)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.92)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.63)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.13)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.45)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.44)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.01)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.27)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.65)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.30)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 0.48)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.65)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded -0.63)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.71)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.06)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.00)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 148
\-------------------------

Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 1.02)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.49)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.60)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.25)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.50)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.54)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.28)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.01)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.34)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.91)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.54)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.59)
36% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 149
\-------------------------

Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.38)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent drove right instead of forward. (rewarded 1.51)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.93)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.31)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.98)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.48)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.50)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.57)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.74)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.77)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.31)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded -0.19)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.72)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.56)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.17)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.46)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 0.41)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 150
\-------------------------

Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.39)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 1.44)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.53)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.62)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 151
\-------------------------

Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . . 
epsilon = 0.0498; alpha = 0.0200

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.78)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.08)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.10)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.09)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.76)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.60)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 2.21)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Testing trial 1
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.43)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.16)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.66)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.07)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.07)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.11)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.10)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.32)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.14)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.14)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.51)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.50)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.50)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.23)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.67)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.16)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.82)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent drove forward instead of right. (rewarded 0.26)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.72)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.72)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.72)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.72)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.72)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.72)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.16)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.30)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.69)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent drove right instead of forward. (rewarded -0.74)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 2
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.13)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.26)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.33)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.48)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.05)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.05)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.05)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.05)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.05)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.66)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.64)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.78)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.58)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.94)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.82)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.49)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Testing trial 3
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.17)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.64)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.68)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.56)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent attempted driving left through a red light. (rewarded -10.58)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent attempted driving left through a red light. (rewarded -10.58)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.14)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.87)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.43)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.39)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.27)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Testing trial 4
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.03)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.80)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.21)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.59)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.17)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.54)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.63)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.15)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.47)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.47)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.47)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.47)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.47)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.47)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.22)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.22)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.22)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.22)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.22)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.14)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.31)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.31)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.31)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.31)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.31)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.00)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.00)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.00)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.00)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.52)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 5
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.34)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.26)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.48)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.38)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.37)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Testing trial 6
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.93)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'left')
Agent followed the waypoint left. (rewarded 1.57)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.97)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.79)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded -0.00)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.10)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.11)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 0.66)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.16)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.97)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded -0.06)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded -0.28)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.09)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.19)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.36)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.01)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.46)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.97)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 7
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.56)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.07)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.38)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.38)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'right')
Agent drove forward instead of left. (rewarded 0.84)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.83)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.96)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.26)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent drove right instead of forward. (rewarded 1.27)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 1.63)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.12)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.53)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 0.88)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.24)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.26)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.72)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.28)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.19)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 8
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.44)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.92)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.32)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.32)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.32)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.32)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.32)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.28)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.51)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.31)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.46)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.65)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.58)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.58)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 0.22)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.86)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.02)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.02)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.09)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.61)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.20)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.47)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.62)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.13)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.47)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.47)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.54)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.37)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.57)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.81)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.65)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.65)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.65)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 9
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.08)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.00)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.00)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.00)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.00)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.00)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.00)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.00)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.01)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.09)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.08)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent drove right instead of left. (rewarded 0.37)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent drove right instead of left. (rewarded 0.37)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.66)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent drove right instead of left. (rewarded 0.02)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.37)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.13)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.41)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 0.42)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 0.42)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 0.42)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.40)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.65)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.07)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.47)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.82)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 10
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.28)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.24)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.31)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.14)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'right', 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.60)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.06)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.06)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.49)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.37)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.05)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.10)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.69)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.10)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.47)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.35)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

Simulation ended. . . 
<matplotlib.figure.Figure at 0x5bbb080>
In [47]:
# Load the 'sim_improved-learning' file from the improved Q-Learning simulation
vs.plot_trials('sim_improved-learning.csv')
print "decay function is :ϵ=e−at, alpha=0.02,epsilon=0.5"
decay function is :ϵ=e−at, alpha=0.02,epsilon=0.5

Question 7

Using the visualization above that was produced from your improved Q-Learning simulation, provide a final analysis and make observations about the improved driving agent like in Question 6. Questions you should answer:

  • What decaying function was used for epsilon (the exploration factor)?
  • Approximately how many training trials were needed for your agent before begining testing?
  • What epsilon-tolerance and alpha (learning rate) did you use? Why did you use them?
  • How much improvement was made with this Q-Learner when compared to the default Q-Learner from the previous section?
  • Would you say that the Q-Learner results show that your driving agent successfully learned an appropriate policy?
  • Are you satisfied with the safety and reliability ratings of the Smartcab?
In [2]:
%run smartcab/agent.py
/-------------------------
| Training trial 1
\-------------------------

Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . . 
epsilon = 1.0000; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.57)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.38)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.32)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 0.98)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.66)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.62)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.66)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.66)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.55)
71% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.55)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.41)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded -0.06)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.74)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.63)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.63)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.63)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.63)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.63)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.63)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.63)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.45)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.26)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.50)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.15)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.79)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.38)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.18)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.47)
34% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.47)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.81)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.81)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.56)
29% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.56)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.95)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.65)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.92)
17% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.92)
17% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.92)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.18)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.62)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.54)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.62)
6% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.62)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.13)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded -0.39)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 2
\-------------------------

Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9980; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.91)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.57)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.06)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.77)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.77)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.77)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.77)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.85)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.06)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.95)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.65)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.89)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.89)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.46)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.46)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.46)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.57)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent drove left instead of right. (rewarded 0.92)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent drove left instead of right. (rewarded 0.92)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent drove left instead of right. (rewarded 0.92)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent drove left instead of right. (rewarded 0.92)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent drove left instead of right. (rewarded 0.92)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.41)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded 0.37)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.31)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.70)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.35)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.35)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.35)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.35)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.14)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.10)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.43)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.82)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.50)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.58)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 3
\-------------------------

Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9960; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.85)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.84)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.76)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.76)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.76)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.76)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.76)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.76)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.87)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.91)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.91)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.18)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.18)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.05)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.22)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.85)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.28)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.78)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.34)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.34)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.34)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.34)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.57)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.72)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.72)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.24)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.63)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.35)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.99)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.99)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.99)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 4
\-------------------------

Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9940; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.62)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.61)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.13)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.64)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.84)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.83)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.41)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.68)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.68)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.68)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.68)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.52)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.14)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.14)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.14)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.95)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.32)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.55)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.55)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.61)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.61)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.27)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.80)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.64)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.10)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.33)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded -0.03)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.71)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.21)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 5
\-------------------------

Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9920; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.99)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.56)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.55)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.52)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.18)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.82)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.28)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.22)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.22)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.88)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.93)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.05)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.17)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.29)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.29)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.29)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 0.47)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.42)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.76)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.01)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 6
\-------------------------

Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9900; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove left instead of right. (rewarded 0.85)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.21)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.71)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.71)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.71)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.71)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.71)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.92)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.24)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.55)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.67)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.40)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.71)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.62)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.75)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.75)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.35)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.86)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.86)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.86)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.86)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded -0.18)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.85)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.85)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.85)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.85)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.48)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.74)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded -0.28)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.41)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.18)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.35)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.17)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', None)
Agent drove right instead of left. (rewarded -0.07)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.90)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.10)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.10)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.10)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.87)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 7
\-------------------------

Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9881; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.74)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.15)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.52)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.68)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.65)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.70)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.70)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.40)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.13)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.13)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.33)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.17)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.27)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.04)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.84)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.44)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.29)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.49)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.49)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.29)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.15)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.99)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.83)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.10)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.01)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.01)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.11)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 8
\-------------------------

Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9861; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.71)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.84)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.71)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.27)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.79)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.44)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.31)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.90)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.97)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.11)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.34)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.94)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.21)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.21)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.21)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.21)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.04)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.04)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.04)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.04)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.29)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.83)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.83)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.83)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.83)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.91)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 9
\-------------------------

Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9841; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.61)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.67)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.42)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.04)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.57)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.56)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.56)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.56)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.57)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.33)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.19)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.75)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.02)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.56)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.85)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.07)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent drove forward instead of right. (rewarded 0.76)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.43)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove left instead of right. (rewarded -0.36)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.49)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.62)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.99)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.09)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 10
\-------------------------

Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9822; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.70)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.31)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.70)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.06)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.84)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.04)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.19)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.13)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.18)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.48)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.74)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.74)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.74)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove forward instead of right. (rewarded 0.12)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.36)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.79)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.57)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.53)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.67)
24% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 11
\-------------------------

Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9802; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.00)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.25)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.80)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.12)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.75)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.26)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.32)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.10)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.68)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.68)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.68)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.33)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.00)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.26)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 0.89)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.78)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.56)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.08)
16% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 12
\-------------------------

Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9782; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.29)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.34)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.88)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.85)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.82)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.70)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.15)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.84)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.84)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.84)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.84)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.10)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.25)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.53)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.27)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.08)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.25)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.45)
36% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 13
\-------------------------

Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9763; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.84)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.65)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.76)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.39)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.05)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.24)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.24)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.24)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.17)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.31)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.97)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.97)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.97)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.97)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.97)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.97)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.52)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.68)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.26)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 14
\-------------------------

Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9743; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.47)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 1.14)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.47)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.47)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.47)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.47)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.18)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.05)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.65)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.07)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.07)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.07)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.07)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.90)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.90)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.42)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.42)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.42)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.51)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.50)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.22)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.31)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.75)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded -0.03)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 0.04)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.47)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.47)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.47)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.47)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.47)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.56)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.18)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.83)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.83)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.83)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.83)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.29)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove right instead of forward. (rewarded 0.24)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 15
\-------------------------

Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9724; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.42)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.15)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.70)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.70)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.70)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.70)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.86)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.95)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.78)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent followed the waypoint forward. (rewarded 0.98)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.37)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.40)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.54)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.92)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.35)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.90)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.13)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.35)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.35)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.11)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.68)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.51)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded -0.13)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded -0.13)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.70)
33% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.70)
33% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.70)
33% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.70)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.24)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.03)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -9.28)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.80)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.80)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.80)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.80)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.57)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.48)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 16
\-------------------------

Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9704; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.86)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.93)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.93)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.61)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.08)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.96)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.43)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.27)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.27)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.27)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.27)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 0.67)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.64)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.64)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.64)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.64)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.64)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.64)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.83)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.22)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.13)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.41)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent drove forward instead of left. (rewarded 0.78)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.80)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 17
\-------------------------

Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9685; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.37)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.44)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.99)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.40)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.16)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove forward instead of right. (rewarded 1.35)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove forward instead of right. (rewarded 1.35)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove forward instead of right. (rewarded 1.35)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.23)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.97)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove forward instead of left. (rewarded 0.51)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.55)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 1.09)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.19)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.84)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.84)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.84)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 0.94)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 0.94)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.28)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.38)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.63)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.52)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.57)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.87)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded -0.25)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.49)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 18
\-------------------------

Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9666; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.63)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.38)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.07)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.61)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.51)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.51)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.78)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.53)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.87)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded -0.09)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.45)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.94)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 0.49)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 0.49)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 0.49)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 0.49)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 0.49)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.94)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.43)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'right')
Agent drove right instead of left. (rewarded 0.08)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 19
\-------------------------

Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9646; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.49)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.11)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.98)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent drove right instead of left. (rewarded 0.32)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.67)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.80)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.54)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.70)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.26)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.60)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.60)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.60)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.60)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.60)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove left instead of right. (rewarded 0.79)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove left instead of right. (rewarded 0.79)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove left instead of right. (rewarded 0.79)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove left instead of right. (rewarded 0.79)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove left instead of right. (rewarded 0.79)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.28)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.28)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.28)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove right instead of left. (rewarded 0.43)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.62)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.62)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.62)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.02)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.88)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.25)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 20
\-------------------------

Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9627; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.44)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.16)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.97)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.36)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.15)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.78)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.94)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.94)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.94)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.94)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.83)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.05)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent drove left instead of right. (rewarded 0.19)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.92)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.92)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.92)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.18)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.02)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent attempted driving left through a red light. (rewarded -9.01)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.37)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded -0.08)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.00)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.00)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.19)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.19)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.19)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.11)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 21
\-------------------------

Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9608; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove left instead of right. (rewarded 1.47)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.28)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.29)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.16)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.74)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.74)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.40)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.69)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.80)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.67)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.05)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.29)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.62)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.08)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.08)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.08)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.08)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.08)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.01)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 22
\-------------------------

Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9589; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.69)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.06)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.06)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.93)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.51)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.51)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.51)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.08)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.42)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.03)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.19)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.41)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.08)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.08)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.13)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.13)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.13)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.61)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.61)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.61)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.61)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.61)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.17)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.73)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.37)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.29)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.29)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.29)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.21)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.73)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.69)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.55)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 23
\-------------------------

Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9570; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.99)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.20)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.29)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.79)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.82)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded 0.75)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.62)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.62)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.99)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.14)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.71)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.20)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.01)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.47)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.28)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.22)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.51)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 1.16)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.54)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.66)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.26)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.84)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.50)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.50)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.50)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.50)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.50)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.39)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.44)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 24
\-------------------------

Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9550; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.66)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.96)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.33)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.73)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.52)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.91)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 0.46)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.03)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.22)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.74)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.53)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.56)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove left instead of forward. (rewarded 0.06)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.44)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove forward instead of left. (rewarded -0.31)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.70)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.18)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.86)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 25
\-------------------------

Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9531; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.37)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.32)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.80)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.48)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.83)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.16)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.46)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.40)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.27)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.57)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.57)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.57)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.57)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.57)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.74)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.19)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.01)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.29)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.29)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.29)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.53)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.37)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.37)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.41)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.85)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.85)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.85)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.85)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.85)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.54)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.00)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 26
\-------------------------

Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9512; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.09)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent drove forward instead of right. (rewarded 1.94)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.36)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.76)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.13)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.13)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.53)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.24)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.20)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.17)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.17)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.17)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.17)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.55)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.45)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.08)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.59)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.59)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.59)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.77)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.44)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 27
\-------------------------

Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9493; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.22)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.48)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.93)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.93)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.78)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded -0.14)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded -0.16)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.94)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.60)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 0.92)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 0.92)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 0.92)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 0.92)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.33)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.56)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.56)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.36)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.36)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.92)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.02)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 28
\-------------------------

Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9474; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.93)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.64)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.99)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.99)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.99)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.99)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.49)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.70)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent drove right instead of left. (rewarded 1.12)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.43)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.43)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.72)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.72)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.72)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.72)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.34)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.93)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.69)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.13)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.13)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded -0.13)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.35)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.29)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.77)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.77)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.04)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.92)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.18)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.14)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.14)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.14)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.14)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.57)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.57)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.12)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.02)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent drove forward instead of right. (rewarded 0.11)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded -0.54)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.61)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 29
\-------------------------

Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9455; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.63)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.82)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.41)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.65)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.80)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.80)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.80)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.76)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent drove forward instead of right. (rewarded 0.37)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.23)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.92)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.48)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.48)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.48)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.48)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.48)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.30)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.94)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.40)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.40)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.40)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.40)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.11)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.03)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 1.64)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 30
\-------------------------

Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9436; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.81)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 0.11)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.36)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.36)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.36)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.22)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.47)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove left instead of forward. (rewarded 1.65)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove left instead of right. (rewarded 0.02)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove left instead of right. (rewarded 0.02)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove left instead of right. (rewarded 0.02)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove left instead of right. (rewarded 0.02)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.31)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.59)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.67)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.67)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.67)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.67)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.67)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.93)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.62)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.62)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.44)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.96)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.74)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.59)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 31
\-------------------------

Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9418; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.74)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.52)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.26)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.15)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.15)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.15)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.31)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.54)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.74)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.75)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.75)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.75)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.75)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.51)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.75)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.49)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.71)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.90)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.64)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.24)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.13)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.27)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.19)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.47)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.58)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.42)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent drove right instead of forward. (rewarded -0.30)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.64)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.29)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 32
\-------------------------

Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9399; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.90)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.96)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.13)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.26)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.40)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.88)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.39)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.97)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.97)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.09)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.00)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove left instead of right. (rewarded 0.41)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.97)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.97)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.97)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.97)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.71)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.30)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 33
\-------------------------

Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9380; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 0.73)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.93)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.56)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.94)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.94)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.94)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.99)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.99)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.99)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.99)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.76)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.43)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.43)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.43)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 2.76)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.21)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.21)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.21)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.21)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.21)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.21)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.14)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.37)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.37)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.47)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.82)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.04)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.42)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.24)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded -0.43)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.19)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.44)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.72)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.93)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 34
\-------------------------

Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9361; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent followed the waypoint left. (rewarded 1.97)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.87)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.04)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.86)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.14)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.14)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.14)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.14)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.13)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.32)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.22)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.34)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.25)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.59)
63% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 35
\-------------------------

Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9343; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.92)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.57)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.00)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.31)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.59)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.87)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.96)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.96)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.38)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.38)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 0.80)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.20)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.44)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.83)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.61)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.32)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.32)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.32)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.32)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.41)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent drove left instead of right. (rewarded 0.65)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.01)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded -0.60)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 36
\-------------------------

Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9324; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.97)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 0.76)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.57)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.57)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.57)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.57)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.57)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.95)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.31)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.90)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.17)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.35)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.35)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.51)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded 0.15)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.48)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.35)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.44)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.44)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.44)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.44)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.44)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.06)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.02)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded -0.59)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.14)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.50)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.27)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 37
\-------------------------

Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9305; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 1.62)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent properly idled at a red light. (rewarded 2.13)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.88)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.35)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.48)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.41)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.41)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.41)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.84)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.79)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent followed the waypoint right. (rewarded 0.83)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.80)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.03)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.46)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.90)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.46)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.00)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.21)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.04)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.04)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.04)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.02)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.68)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 38
\-------------------------

Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9287; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.32)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove left instead of forward. (rewarded 1.40)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove left instead of forward. (rewarded 1.40)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove left instead of forward. (rewarded 1.40)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.17)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.17)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.32)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.32)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.05)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.14)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove forward instead of right. (rewarded 1.62)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.44)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.96)
63% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 39
\-------------------------

Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9268; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.15)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.33)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.65)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.88)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.88)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.92)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.51)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.02)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.62)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.17)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.03)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded -0.23)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded -0.65)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.50)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.51)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 40
\-------------------------

Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9250; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.41)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.50)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.59)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.02)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.02)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.96)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.96)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.92)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.51)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.51)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.20)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.78)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.50)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.50)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.19)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.40)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.18)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.35)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.35)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.35)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.27)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.08)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 41
\-------------------------

Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9231; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.02)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.62)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.62)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.62)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.18)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.84)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.84)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.84)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.84)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.89)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.93)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.42)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.52)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.98)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.30)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.75)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.35)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.00)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove forward instead of right. (rewarded 0.38)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.55)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.73)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.52)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.73)
8% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 42
\-------------------------

Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9213; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.76)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.95)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.50)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.80)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.25)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.61)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.08)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.74)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded -0.04)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded -0.04)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded -0.04)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded -0.04)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.29)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.03)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.46)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.09)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.09)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.09)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.09)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.09)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.05)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.00)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 43
\-------------------------

Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9194; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.52)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.15)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.01)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.72)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.88)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 0.88)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.88)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.49)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.56)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.56)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.56)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.56)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.13)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.13)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.93)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.59)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.59)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.59)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.59)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.09)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.51)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.93)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent drove forward instead of right. (rewarded 0.12)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.37)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 0.23)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 44
\-------------------------

Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9176; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.56)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.70)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.49)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.11)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.35)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.87)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.84)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.84)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.84)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.84)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.40)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent drove right instead of left. (rewarded 1.01)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.24)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 0.82)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.79)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.21)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.02)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.81)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.81)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.52)
24% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 45
\-------------------------

Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9158; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.94)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.39)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.57)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.57)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.57)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.57)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.57)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.57)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.09)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 1.16)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.19)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.42)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.56)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.69)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.01)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.93)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.69)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 0.71)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.17)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.17)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.17)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.17)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 0.86)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 0.86)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 0.86)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 0.86)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.57)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.26)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 46
\-------------------------

Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9139; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.54)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.19)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.88)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.25)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.38)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.44)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent followed the waypoint right. (rewarded 1.02)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.64)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.10)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 1.14)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.07)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.07)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.07)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.07)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.53)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.45)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.45)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.45)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.45)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.35)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.90)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.91)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.49)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded 1.44)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.97)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.23)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.48)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.53)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 47
\-------------------------

Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9121; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 1.80)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.69)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.54)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.94)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.94)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.80)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 1.90)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.02)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 1.49)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.59)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.23)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.29)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.41)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.69)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.15)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.64)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.45)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.45)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.45)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.04)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.70)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.28)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.28)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.10)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.75)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.00)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.94)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.59)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.24)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded -0.53)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded -0.14)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.68)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 48
\-------------------------

Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9103; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.40)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.74)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.63)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.19)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.19)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.19)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.19)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.26)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.06)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.06)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.06)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.35)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.88)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.28)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.05)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.79)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.15)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.76)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent drove forward instead of right. (rewarded -0.01)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.89)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.91)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.71)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.11)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.64)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.35)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.15)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 49
\-------------------------

Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9085; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.21)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.88)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.62)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.37)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.41)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.25)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.37)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.05)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.05)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.20)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.49)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 1.00)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'right')
Agent attempted driving left through a red light. (rewarded -10.93)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.69)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.92)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
5% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 50
\-------------------------

Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9066; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.79)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.39)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.81)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.15)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.79)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.03)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.00)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.58)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.85)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.15)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.10)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.40)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.18)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 0.41)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 51
\-------------------------

Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9048; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.44)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.04)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.29)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.94)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.94)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.94)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.94)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.94)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.94)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.20)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.68)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.74)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.59)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 52
\-------------------------

Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9030; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.53)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.11)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.94)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.06)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.72)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.75)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.53)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.53)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.53)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.53)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.53)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.15)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.74)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.70)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.60)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.41)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.41)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.41)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.41)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.77)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.77)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 53
\-------------------------

Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.9012; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.40)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.53)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.50)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.02)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.69)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.08)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.08)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.08)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.08)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.08)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.08)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.08)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.71)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.46)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.63)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.76)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.65)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent followed the waypoint left. (rewarded 0.68)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.69)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.26)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.95)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.48)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 54
\-------------------------

Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8994; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.00)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.30)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.80)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.26)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 1.48)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.92)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent drove left instead of right. (rewarded 1.85)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.72)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded 1.04)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.30)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.10)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent drove right instead of left. (rewarded 1.53)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.71)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded -0.19)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded -0.19)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded -0.19)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded -0.19)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent drove forward instead of right. (rewarded 0.12)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.53)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.21)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.21)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.21)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.02)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.70)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.30)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.88)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.41)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 55
\-------------------------

Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8976; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.58)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.99)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded -0.02)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.22)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.22)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.32)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.36)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.37)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.06)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.06)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.06)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.26)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.97)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.78)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.66)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.29)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.84)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.20)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.39)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 56
\-------------------------

Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8958; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.70)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.25)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.43)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.43)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.80)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.80)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.80)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.14)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.14)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.47)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.77)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded -0.01)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.91)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.14)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.95)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.11)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.48)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.48)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.24)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.64)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.77)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 57
\-------------------------

Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8940; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.85)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.00)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.11)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.61)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.56)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.85)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.63)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.63)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.63)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.63)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.63)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded -0.04)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent drove forward instead of right. (rewarded 0.18)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 0.74)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.64)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.38)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.96)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.46)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.46)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.83)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.18)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.62)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded -0.12)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 58
\-------------------------

Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8923; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.07)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', None)
Agent drove forward instead of right. (rewarded 0.55)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent drove forward instead of right. (rewarded 1.01)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.43)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 59
\-------------------------

Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8905; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.39)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.85)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.68)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.21)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.09)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.93)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.30)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.30)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.30)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.30)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded -0.09)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded -0.09)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded -0.09)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.87)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.51)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.10)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.66)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 2.49)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.87)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.87)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.96)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.09)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.83)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.83)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.83)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.83)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.95)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded -0.26)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded -0.26)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded -0.26)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded -0.26)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded -0.26)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded -0.39)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.37)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.25)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 60
\-------------------------

Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8887; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.63)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent drove forward instead of right. (rewarded 0.13)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.34)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.79)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.94)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.94)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.94)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.94)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.88)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.88)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.39)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.30)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.02)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.02)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.02)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.02)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.02)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.65)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.16)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.03)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.31)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 0.92)
36% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 61
\-------------------------

Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8869; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.64)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.07)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.89)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.01)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.14)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.14)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.14)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.14)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.26)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.54)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.39)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.37)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.92)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.81)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.66)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 1.12)
48% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 1.12)
48% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 1.12)
48% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 1.12)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.00)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.49)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.09)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.09)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.52)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.50)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.24)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.24)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.24)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.24)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.24)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.24)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.48)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.10)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.10)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded -0.29)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.37)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 62
\-------------------------

Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8851; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.39)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.70)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.97)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.97)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.97)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.97)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.97)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.79)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.24)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.62)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.62)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.62)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.62)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.55)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.30)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.34)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.68)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.68)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.68)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.54)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.81)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.81)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.17)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.12)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.64)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.57)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.57)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.57)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.57)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.57)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.40)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.88)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.46)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.53)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.76)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.76)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.76)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.76)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.84)
8% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 63
\-------------------------

Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8834; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 1.85)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.62)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.52)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.27)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.09)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.69)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.74)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.35)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.29)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.54)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove forward instead of left. (rewarded -0.14)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.48)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.11)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.61)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.61)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.61)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 1.34)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.67)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.67)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.67)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.07)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.07)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.15)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 64
\-------------------------

Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8816; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.72)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.60)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.60)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.41)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.41)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.41)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.41)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.38)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.53)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.53)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.53)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.53)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.53)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.22)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.83)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.83)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.83)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.83)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.83)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.51)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.65)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 65
\-------------------------

Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8799; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.62)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.65)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.17)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.43)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.82)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.82)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.82)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.82)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.85)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.66)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.66)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.66)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.66)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.66)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.53)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.62)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.14)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.51)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.51)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.16)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 1.22)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.24)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.24)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.24)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.42)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.03)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.03)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded -0.61)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'right')
Agent followed the waypoint right. (rewarded 1.16)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.03)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.79)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 66
\-------------------------

Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8781; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.39)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.52)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.14)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.81)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.37)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.87)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.87)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.87)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.87)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.87)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.87)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.29)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.26)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.41)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.94)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.42)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.32)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.16)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.30)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.87)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.06)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.57)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.63)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.38)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 67
\-------------------------

Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8763; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.83)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.93)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.70)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.70)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.70)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.70)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.70)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.70)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.70)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.70)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.94)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.52)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.04)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.76)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.72)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 68
\-------------------------

Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8746; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.00)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.74)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.40)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.82)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.21)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.21)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.21)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.29)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.16)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.69)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.64)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.08)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.08)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.08)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.23)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.64)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.25)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.25)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.25)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.25)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.25)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.35)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.17)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.01)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.78)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.78)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.78)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.07)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.34)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.13)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.03)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 69
\-------------------------

Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8728; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.42)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.53)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.14)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.03)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.25)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.48)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.43)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.75)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.21)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.75)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.28)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.18)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.18)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.18)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.63)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.53)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.50)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.53)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.53)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.53)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.79)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove forward instead of left. (rewarded -0.58)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove forward instead of left. (rewarded -0.58)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove forward instead of left. (rewarded -0.58)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.13)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded -0.46)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 70
\-------------------------

Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8711; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.25)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.84)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.53)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.84)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.78)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.78)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.78)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.45)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.14)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.13)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.68)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.74)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.74)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.88)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.53)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.13)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.72)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.45)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.09)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 71
\-------------------------

Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8694; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.14)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.75)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.58)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.26)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.43)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.16)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.16)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.16)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.91)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent followed the waypoint left. (rewarded 1.58)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.86)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.66)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.11)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.24)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.24)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.24)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.24)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.24)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.24)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.24)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded -0.09)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove left instead of right. (rewarded 1.19)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded -0.67)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.46)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.25)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 72
\-------------------------

Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8676; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.44)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.48)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.11)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.43)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.42)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.92)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 2.36)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.34)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.34)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.34)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.34)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.34)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent drove right instead of left. (rewarded 1.85)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 1.72)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.35)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.81)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.51)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.51)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.51)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.51)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.66)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.40)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.13)
49% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.13)
49% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.13)
49% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.13)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.15)
46% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.15)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.06)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.57)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.17)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.92)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
29% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 73
\-------------------------

Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8659; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.37)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.05)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.90)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.87)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.29)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.84)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded -0.10)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.57)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.83)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 74
\-------------------------

Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8642; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.51)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.18)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.84)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.73)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.73)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.73)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.73)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.16)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.16)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.86)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.86)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.86)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.86)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.13)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.15)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.35)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.14)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.58)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.91)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.32)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.43)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.32)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.46)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.04)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.51)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.44)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.01)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.37)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 75
\-------------------------

Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8624; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.00)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.52)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.46)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.82)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.36)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.75)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.91)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.32)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.32)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.27)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.27)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.27)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.35)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.31)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.31)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.90)
36% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 76
\-------------------------

Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8607; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded 0.20)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.74)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.68)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.81)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.92)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.96)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.19)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.29)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.29)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.29)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.29)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.27)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.39)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.39)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.39)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent drove right instead of left. (rewarded 1.42)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.75)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.75)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.75)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.75)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.19)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.51)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.64)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.14)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.17)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.27)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.75)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.05)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.31)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.50)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.72)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.37)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.77)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded -0.22)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.65)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 77
\-------------------------

Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8590; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.74)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.83)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.02)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.64)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.12)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.12)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.54)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.75)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.33)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.31)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.70)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.35)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.35)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.90)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.92)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.04)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 78
\-------------------------

Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8573; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent drove left instead of right. (rewarded 0.90)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.53)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.09)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.02)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.08)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.68)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.28)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.90)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.34)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.00)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.39)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.29)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.73)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.62)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.62)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.63)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.63)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.91)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.53)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.94)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.29)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.39)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded -0.25)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.89)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.82)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.02)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.27)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.10)
7% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.10)
7% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.10)
7% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.10)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.36)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.38)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 79
\-------------------------

Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8556; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.45)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.12)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.33)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.45)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.81)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.81)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 0.96)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.21)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.21)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.21)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.03)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.05)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.75)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.64)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.74)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.74)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.74)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.74)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.74)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.35)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.18)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.86)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.48)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.48)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.48)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.48)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.37)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.50)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.64)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.60)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.96)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.74)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.74)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.74)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.74)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.74)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.01)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 80
\-------------------------

Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8538; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.45)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.65)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.31)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.70)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.65)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.80)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.20)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.20)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.20)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.94)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.29)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.43)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.43)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.43)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.43)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded -0.02)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.14)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.73)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.77)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.51)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.51)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.51)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.51)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.51)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.10)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.37)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.79)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.99)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.73)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 1.51)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.00)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.78)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.77)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.77)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.18)
13% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.18)
13% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.18)
13% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.18)
13% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.18)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.63)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.35)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.01)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 81
\-------------------------

Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8521; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.33)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.72)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.88)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.88)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.88)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.88)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.88)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.96)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.72)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent drove right instead of left. (rewarded 0.18)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.00)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.14)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.21)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.72)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.64)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.77)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.39)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.39)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.39)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.77)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.04)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.04)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.04)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.04)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.04)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.44)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.16)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.02)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.48)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.02)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.84)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.87)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 82
\-------------------------

Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8504; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.05)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.37)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 1.53)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.88)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.66)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.39)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.86)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.86)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.86)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.18)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.18)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.18)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.85)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.85)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.34)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.89)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.44)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.05)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.43)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.77)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.77)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.92)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.92)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.41)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.39)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.46)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.46)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.46)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.26)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.08)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.71)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.01)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.51)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 83
\-------------------------

Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8487; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.63)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.43)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.23)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.84)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.67)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.67)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.67)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.67)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.67)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.14)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.71)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.75)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.75)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.78)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.13)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.13)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.59)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.07)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.18)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.97)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.73)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.62)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.50)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded 0.23)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 84
\-------------------------

Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8470; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.50)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.89)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.59)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.50)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.29)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.85)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.85)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.85)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.85)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.65)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -41.00)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.09)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.38)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.38)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.38)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded -0.14)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.54)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.54)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.30)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.30)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.30)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.30)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.30)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.66)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.63)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.94)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 85
\-------------------------

Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8454; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.30)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.92)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.49)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.77)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.74)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.48)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.06)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.26)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.63)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.88)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.89)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.83)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.96)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 1.30)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.92)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.60)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.34)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.34)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.34)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.34)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.34)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.34)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.82)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.49)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 86
\-------------------------

Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8437; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.33)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.71)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.04)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.76)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.99)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.03)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.46)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.35)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.35)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.95)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.79)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.79)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.79)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.79)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded -0.23)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.71)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.71)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded -0.17)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 0.41)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.78)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.78)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.78)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.78)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded -0.11)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.65)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 87
\-------------------------

Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8420; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.81)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.04)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.04)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.04)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.75)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.80)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.49)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 1.15)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.66)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.69)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.69)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.69)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.69)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.63)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.57)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.38)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.04)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.04)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.15)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.10)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.89)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 88
\-------------------------

Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8403; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.58)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.98)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.98)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.98)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.58)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.00)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.38)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.84)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.39)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.98)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.64)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.66)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.54)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded -0.39)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.96)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.65)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 89
\-------------------------

Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8386; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.70)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.73)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.22)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.47)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.10)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.57)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.57)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.57)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.57)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.57)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.04)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.17)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.34)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.73)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.31)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.31)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.31)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.31)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.26)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.49)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.30)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.98)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.07)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.07)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.72)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 90
\-------------------------

Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8369; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.49)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.99)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.60)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 1.74)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.33)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.61)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 1.84)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.21)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.61)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.61)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.61)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.63)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.88)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.49)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.24)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.67)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.26)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.62)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.62)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.69)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.69)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.69)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.69)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.69)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.47)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.12)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.11)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 91
\-------------------------

Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8353; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.73)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.61)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.62)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.47)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.86)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.72)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 1.83)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.68)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.22)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.22)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.22)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.12)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.13)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.17)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.10)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.40)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.40)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.47)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.92)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.20)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 0.11)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 0.11)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 0.11)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.53)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 92
\-------------------------

Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8336; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.27)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.75)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.51)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.51)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.51)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.81)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.91)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.91)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.91)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.91)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.82)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.73)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.80)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.80)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.80)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.47)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.47)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.77)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.77)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.77)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 0.18)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.31)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.32)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.87)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.65)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.75)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.53)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.53)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.53)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.53)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.29)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded -0.46)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 93
\-------------------------

Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8319; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.37)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove left instead of right. (rewarded 0.56)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.57)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.27)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.54)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.37)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.95)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.78)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.03)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.03)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.03)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.17)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 1.72)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.61)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.61)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.52)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.61)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.61)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.61)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.18)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.57)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.71)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.31)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.38)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 94
\-------------------------

Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8303; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.76)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.43)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.28)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.17)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.17)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.17)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.52)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.43)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.33)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.58)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.58)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.22)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.39)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.56)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.56)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.56)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.13)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.79)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove left instead of right. (rewarded 0.74)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.44)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.24)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.17)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'right')
Agent properly idled at a red light. (rewarded 0.44)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.05)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.05)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.05)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.05)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 95
\-------------------------

Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8286; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.40)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.12)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.11)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.36)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.35)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.70)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.70)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.70)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.81)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.05)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.05)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.05)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.05)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.18)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.34)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.34)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.34)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.34)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.34)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove left instead of forward. (rewarded 0.85)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.76)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.76)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.76)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.24)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.01)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.58)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.01)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.12)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded -0.10)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.50)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 96
\-------------------------

Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8270; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.98)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.37)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.02)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.14)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.04)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.17)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.22)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.84)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.84)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.84)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.84)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.63)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.51)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.11)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.11)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.11)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.91)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.78)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.33)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.23)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.54)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.11)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.45)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 97
\-------------------------

Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8253; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.86)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.16)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.90)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 1.23)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.75)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.28)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.35)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 1.13)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.24)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.52)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.11)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.68)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.55)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.55)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.55)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.55)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.60)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent drove forward instead of right. (rewarded -0.20)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 0.88)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.14)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 98
\-------------------------

Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8237; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.85)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.40)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.15)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.83)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.83)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.83)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.39)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 1.22)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.44)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.44)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.44)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.30)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.76)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.08)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.34)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.34)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.31)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.31)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.14)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.94)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.37)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.37)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.37)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.37)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.32)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.83)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
27% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 99
\-------------------------

Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8220; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.04)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.31)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.84)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.53)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.53)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.53)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.63)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.63)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.42)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.32)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.70)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 100
\-------------------------

Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8204; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.77)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.44)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.84)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.09)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.91)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.91)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.91)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.91)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.22)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.54)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.79)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.33)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded 0.77)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded 0.77)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded 0.77)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.64)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.42)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.23)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.99)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.87)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.80)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.11)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.11)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.11)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.11)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove left instead of right. (rewarded -0.08)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.12)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.12)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.12)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.28)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.28)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.15)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'right')
Agent attempted driving left through a red light. (rewarded -9.80)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.06)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 101
\-------------------------

Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8187; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.02)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded 1.61)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.90)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 0.03)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.52)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.47)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.99)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.99)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.99)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.10)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.19)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.85)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.85)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.66)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.95)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.88)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.35)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 102
\-------------------------

Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8171; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.58)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.95)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.64)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.34)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.39)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.39)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.30)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.50)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.63)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.26)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.42)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.27)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.79)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.14)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.35)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.67)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.64)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.01)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 103
\-------------------------

Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8155; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.18)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.80)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.81)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.27)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.23)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.15)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.42)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 104
\-------------------------

Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8138; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.60)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.50)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.97)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.94)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.46)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.46)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.46)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.46)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.46)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.46)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.05)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.54)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.54)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.54)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.54)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.54)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.42)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.29)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.72)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.72)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.72)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded -0.14)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.58)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.42)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.54)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.07)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.60)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.81)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.81)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.77)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 105
\-------------------------

Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8122; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.45)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.54)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.09)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 0.66)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.27)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.22)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded 0.15)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.09)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.19)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.19)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.19)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.19)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.19)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.19)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove left instead of right. (rewarded 0.85)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.70)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.16)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.64)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.58)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.52)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.06)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.22)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.57)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.62)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.06)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.49)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 106
\-------------------------

Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8106; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.82)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.52)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.52)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.52)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.52)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.52)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.52)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.52)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.39)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.70)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.28)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.86)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.95)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.19)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.77)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.77)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.77)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.06)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.25)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.25)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.25)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.25)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.25)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.67)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.66)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.81)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.89)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.89)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.16)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded -0.02)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.87)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.58)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.76)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.37)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', None)
Agent drove forward instead of right. (rewarded 0.92)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.50)
4% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 107
\-------------------------

Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8090; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.39)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded 0.09)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.40)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.03)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.03)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.23)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.01)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.04)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.28)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.21)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.97)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.07)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.26)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.97)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.31)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.29)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded -0.05)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.39)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 0.99)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.04)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.38)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.40)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.40)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.25)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 108
\-------------------------

Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8073; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.54)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'right')
Agent drove right instead of left. (rewarded 0.22)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.64)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.11)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.64)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.54)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.07)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.94)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 0.44)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.29)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.29)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.29)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.29)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.29)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.95)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.11)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.47)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.08)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.43)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.21)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 109
\-------------------------

Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8057; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.28)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.39)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.47)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.36)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.20)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.96)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.96)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.05)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.05)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.05)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.76)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.19)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.53)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.17)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.87)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.07)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.07)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.07)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.07)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.50)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.20)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.87)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.32)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 110
\-------------------------

Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8041; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.86)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.02)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.02)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', None)
Agent drove forward instead of right. (rewarded 0.03)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'left', None)
Agent drove forward instead of right. (rewarded 0.03)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'left', None)
Agent drove forward instead of right. (rewarded 0.03)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.86)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.06)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.89)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent drove right instead of left. (rewarded 1.16)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent drove right instead of left. (rewarded 1.16)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.56)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.56)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.56)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.28)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.74)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.02)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.73)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove left instead of forward. (rewarded 0.66)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.96)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.53)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent drove forward instead of right. (rewarded -0.46)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.18)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 111
\-------------------------

Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8025; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.05)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.96)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.60)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.45)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.21)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.86)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.96)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.50)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 1.38)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.56)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.56)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.56)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 0.22)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.88)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.96)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove forward instead of left. (rewarded 1.32)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.14)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.92)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.85)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.98)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 112
\-------------------------

Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.8009; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.21)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.84)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 1.53)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.16)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 1.75)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.49)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded 0.83)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.46)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.57)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 113
\-------------------------

Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7993; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.98)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.01)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.66)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.66)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.66)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.66)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.66)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.66)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.66)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.66)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.66)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.22)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.02)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.76)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.76)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.27)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.77)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.54)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.21)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.21)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.21)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.93)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.39)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.39)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.39)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.39)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.69)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.87)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.87)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.87)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.87)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.60)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.12)
27% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 114
\-------------------------

Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7977; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.15)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.74)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.59)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.59)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.59)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.59)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.59)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.05)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.84)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.84)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.84)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.84)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.65)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.06)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.51)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.42)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.31)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.87)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.26)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.15)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 0.88)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.36)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.25)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.98)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'right')
Agent drove right instead of forward. (rewarded 0.29)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.65)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.43)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.28)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.57)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.57)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.57)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.57)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.89)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.42)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.42)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.42)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.32)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.32)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.32)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.92)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.09)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 115
\-------------------------

Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7961; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.51)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.83)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.75)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.20)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.20)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.20)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.83)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 0.99)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.31)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.16)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.27)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.87)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.34)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.02)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.06)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.03)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.24)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.63)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.07)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.36)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded 0.45)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.34)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.75)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.26)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.58)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.54)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.93)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 116
\-------------------------

Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7945; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.32)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 0.85)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.28)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.95)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.95)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.95)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.95)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.95)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.90)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.95)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.34)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.34)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.34)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.60)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.87)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.79)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.79)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.79)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.79)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.52)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.19)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.19)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.56)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 117
\-------------------------

Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7929; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.39)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.74)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.29)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -6.00)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.65)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.23)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.03)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.06)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded -0.08)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.63)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.46)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.16)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.77)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.42)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.80)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.80)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.80)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.80)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.80)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.80)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.80)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.12)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.14)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 0.91)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.53)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.91)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.91)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.53)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent drove forward instead of right. (rewarded 0.97)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.21)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.77)
29% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 118
\-------------------------

Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7914; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.26)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.67)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.25)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.77)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.42)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.73)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.02)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 1.43)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 1.39)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.57)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.71)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.24)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.72)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.43)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.68)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.13)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.60)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.24)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.37)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.40)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.25)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.96)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.06)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.50)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.28)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.45)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.33)
14% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.33)
14% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.33)
14% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.33)
14% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.33)
14% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.33)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 0.94)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.72)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.55)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent drove forward instead of left. (rewarded -0.58)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.31)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 119
\-------------------------

Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7898; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.19)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.07)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.07)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.07)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.07)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.07)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.86)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.65)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.33)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.17)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.17)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.17)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.17)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.57)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded 0.15)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.64)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.64)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.64)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.64)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.84)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.14)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.43)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded -0.00)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded -0.00)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded -0.00)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded -0.00)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded -0.00)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.29)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.54)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.03)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 0.61)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.35)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.67)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 0.98)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.22)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.22)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.22)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.22)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.61)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.33)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.26)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 120
\-------------------------

Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7882; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent followed the waypoint forward. (rewarded 2.23)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.97)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.09)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.59)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.25)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.50)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.63)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 0.51)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 0.51)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 0.51)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 0.51)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.10)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.54)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.42)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.92)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.92)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.92)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.92)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.92)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.92)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'right')
Agent drove right instead of forward. (rewarded 0.17)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.17)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.37)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.38)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.38)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.99)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.84)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.04)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 121
\-------------------------

Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7866; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.27)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.22)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.56)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded -0.01)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.84)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.84)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.84)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.11)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.11)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.68)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.72)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.77)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.22)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded -0.08)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.43)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.80)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove left instead of right. (rewarded -0.09)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.40)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.40)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.40)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.40)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.06)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.11)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.11)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.11)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.11)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.48)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.36)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 122
\-------------------------

Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7851; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.43)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.18)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.56)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.69)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.69)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.19)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.48)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.19)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded -0.09)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.44)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.49)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent drove left instead of forward. (rewarded -0.21)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded -0.16)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.48)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.25)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.42)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.02)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.95)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.35)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.26)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.26)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.18)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.68)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 0.77)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 0.77)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.09)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.09)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.90)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 123
\-------------------------

Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7835; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.01)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.89)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.38)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.71)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.72)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.33)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 124
\-------------------------

Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7819; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 1.23)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.04)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.01)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.36)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.86)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.77)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.52)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.12)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.80)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.94)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.63)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.90)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.27)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.42)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.53)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.53)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.18)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.18)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.18)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.39)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 125
\-------------------------

Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7804; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.06)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.63)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.31)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.67)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove left instead of forward. (rewarded 0.80)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded -0.06)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.02)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove forward instead of right. (rewarded 0.22)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.01)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.34)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.12)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.94)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.94)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.94)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.40)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.40)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.75)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.75)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.75)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.31)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.31)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.21)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.69)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.10)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 126
\-------------------------

Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7788; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.07)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.87)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.71)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.71)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.71)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.71)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.71)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.71)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.49)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.14)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.44)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.44)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.44)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.44)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.44)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.98)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.22)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.02)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.37)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.84)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.25)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.46)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.14)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 0.69)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 127
\-------------------------

Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7772; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.66)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.51)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.92)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.96)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.12)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.12)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.41)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.28)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.83)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.97)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.98)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.94)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.41)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.41)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.41)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.56)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.32)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.51)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.51)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.51)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.12)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.92)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.25)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.19)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.71)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.79)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 128
\-------------------------

Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7757; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.34)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.45)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.45)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.07)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.51)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.51)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.51)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.17)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.61)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.17)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.57)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.43)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.15)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.15)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.69)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.69)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.69)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.01)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.45)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.71)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.23)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.23)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.23)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.23)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.18)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 129
\-------------------------

Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7741; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.28)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.66)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.85)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.85)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.32)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.18)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.98)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.66)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.66)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.66)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.66)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.66)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.13)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.13)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.13)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.47)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.10)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.57)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.30)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove left instead of forward. (rewarded 1.39)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded -0.44)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.02)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.00)
5% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 130
\-------------------------

Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7726; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.67)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.48)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.48)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.48)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.48)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.14)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.45)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.37)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.37)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.37)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.37)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.37)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.14)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.90)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.18)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.10)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.92)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.26)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.28)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.80)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.88)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.56)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.72)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.72)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.72)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.72)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.72)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.87)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.79)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 131
\-------------------------

Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7711; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.84)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove forward instead of right. (rewarded 0.12)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.26)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.23)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 1.68)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.33)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.33)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.33)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.66)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.41)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.96)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.19)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.39)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.84)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.84)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.84)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.83)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.50)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.81)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.23)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.20)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.34)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.77)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.77)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.77)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.77)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.77)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.24)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 132
\-------------------------

Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7695; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.88)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.69)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.57)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.36)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.36)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.41)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove forward instead of left. (rewarded 1.66)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.31)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.33)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.35)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.34)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.56)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.05)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.52)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
24% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 133
\-------------------------

Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7680; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.58)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.65)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.61)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.93)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.67)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.24)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent drove forward instead of right. (rewarded 0.30)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.10)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'right')
Agent drove right instead of left. (rewarded 0.17)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.84)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.84)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.75)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.87)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.08)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.71)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.71)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.71)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.63)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.78)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.13)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.61)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 1.19)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.81)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.30)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.88)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 134
\-------------------------

Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7664; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.16)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.86)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.50)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.50)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.33)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.53)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.78)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.44)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.50)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.36)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.21)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.34)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.26)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.88)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 135
\-------------------------

Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7649; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.92)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.16)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.16)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.16)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.16)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.16)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.43)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.46)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent drove forward instead of right. (rewarded 0.23)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.02)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.60)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.99)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.17)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.31)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.78)
48% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.78)
48% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.78)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.02)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.19)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.19)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.33)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.33)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.82)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.98)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.12)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.12)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.12)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded -0.32)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 1.60)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.72)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 136
\-------------------------

Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7634; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.72)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.06)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.06)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.73)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.66)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.31)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.68)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.68)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.68)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.68)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.68)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.01)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.56)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.92)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.62)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.62)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.40)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.06)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.94)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.45)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.76)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.37)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.52)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.64)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.09)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.09)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.09)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.09)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded -0.14)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.38)
13% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.38)
13% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.38)
13% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.38)
13% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.38)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.65)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.35)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.10)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded -0.74)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 137
\-------------------------

Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7619; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.93)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.55)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.74)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.74)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.75)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.24)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.24)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.93)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.79)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.79)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.61)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.06)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.06)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.06)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.06)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.06)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.52)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.72)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.72)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.72)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.72)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.72)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.16)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.32)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.32)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.32)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.67)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.67)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.89)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.89)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.28)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.30)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.49)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.29)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.47)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.57)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.51)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.82)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.95)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded -0.29)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded -0.78)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.04)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 138
\-------------------------

Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7603; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.88)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.99)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.19)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.19)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.19)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.19)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.19)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.19)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.79)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.13)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.31)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.77)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.13)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.37)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.37)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.03)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.46)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.75)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove right instead of left. (rewarded -0.25)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove right instead of left. (rewarded -0.25)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove right instead of left. (rewarded -0.25)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove right instead of left. (rewarded -0.25)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.55)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.25)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.25)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.25)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.25)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.82)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.19)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.31)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.31)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.31)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.31)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.31)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.32)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.66)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.66)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.66)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.66)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.66)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.66)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.48)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.76)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 139
\-------------------------

Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7588; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.40)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.65)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.60)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.67)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 140
\-------------------------

Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7573; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.65)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.05)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.19)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.19)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.19)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.19)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.94)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.75)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.75)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.75)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.36)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.54)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.72)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.61)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.61)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.61)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.61)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.27)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.80)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.98)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.17)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded -0.15)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.24)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.31)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.93)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.93)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.93)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.93)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded 1.10)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 141
\-------------------------

Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7558; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.50)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.13)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.80)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.43)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.45)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.10)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.10)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.10)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.10)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.10)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.81)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.64)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.64)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.66)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.22)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.58)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.62)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.31)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.22)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.22)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.04)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.60)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.60)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.60)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.62)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.87)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.10)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.07)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.33)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.81)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded -0.29)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 142
\-------------------------

Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7543; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.14)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.36)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.36)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.36)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.36)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.36)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.36)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.36)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.47)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.22)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.25)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.29)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.29)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.29)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.76)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.76)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.72)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.25)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.25)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.25)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.25)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.25)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.40)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.61)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.61)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.15)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.15)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.15)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.65)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.42)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.42)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.00)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.01)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.73)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.81)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.81)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.81)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.05)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 0.21)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 143
\-------------------------

Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7528; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.44)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.61)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent drove forward instead of left. (rewarded 0.52)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.61)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.36)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.54)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.74)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent drove right instead of forward. (rewarded 1.35)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent drove right instead of left. (rewarded 0.71)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.43)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.54)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.07)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.07)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.07)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.05)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.05)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.87)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.87)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.01)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.82)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.35)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.70)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.70)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.70)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.70)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.98)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 1.34)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.18)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.87)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 144
\-------------------------

Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7513; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.89)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.64)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.75)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.32)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.88)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.23)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.05)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.05)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.61)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.28)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.69)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.69)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.69)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.42)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.55)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.69)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.18)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded 1.34)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded 1.34)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded 1.34)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 1.38)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.54)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.58)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.48)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove right instead of left. (rewarded -0.63)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.00)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 145
\-------------------------

Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7498; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.95)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent drove right instead of left. (rewarded 0.17)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.58)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.26)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.40)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.40)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.90)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.41)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.92)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.55)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'right')
Agent drove left instead of forward. (rewarded 1.10)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.25)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.35)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded -0.13)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.32)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.12)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.06)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.35)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.60)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 0.52)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 146
\-------------------------

Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7483; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.43)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.55)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.67)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.92)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.92)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.92)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.01)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.81)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'right')
Agent properly idled at a red light. (rewarded 1.33)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.43)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.43)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.43)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.51)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.48)
57% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 147
\-------------------------

Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7468; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.87)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.21)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.27)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 1.60)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent drove right instead of left. (rewarded 1.36)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent drove right instead of left. (rewarded 1.36)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.58)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.58)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.69)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.09)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.27)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.27)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.82)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.82)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.82)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.82)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.90)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.04)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.66)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.66)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.21)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.57)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 0.71)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 0.68)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.03)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 148
\-------------------------

Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7453; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.42)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.34)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 0.97)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.82)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.61)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.15)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.42)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.81)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.29)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.49)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.43)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.73)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.10)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.03)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 0.80)
5% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 149
\-------------------------

Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7438; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.23)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.12)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.90)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.38)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.38)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.86)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.89)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.35)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.40)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.01)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.68)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 150
\-------------------------

Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7423; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.50)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.67)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.29)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.61)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.44)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.98)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.76)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.71)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.26)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.03)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 0.18)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.85)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.85)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.85)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.85)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.53)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.02)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.37)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.53)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.66)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.66)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.66)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.66)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.66)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.01)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded -0.07)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded -0.07)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded -0.07)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded -0.07)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.50)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.57)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.23)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.31)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.71)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 151
\-------------------------

Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7408; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.43)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.58)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.59)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.50)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.79)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.11)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.73)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.55)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.21)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.56)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.50)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.44)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded 1.65)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.67)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.67)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.67)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.59)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove left instead of right. (rewarded 1.32)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.07)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.81)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.99)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.74)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.53)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.99)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 152
\-------------------------

Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7393; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.39)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.65)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.41)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.80)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.80)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.80)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.80)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.58)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.31)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.49)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.49)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.30)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.30)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.30)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.30)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.99)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.99)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.43)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.43)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.43)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.43)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 0.07)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.41)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.41)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.41)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.28)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.73)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.73)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.73)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.73)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.73)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.12)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.09)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.02)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.02)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.02)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.02)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.42)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 153
\-------------------------

Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7379; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.53)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.24)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent drove forward instead of right. (rewarded 1.23)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.40)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.29)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 1.71)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.13)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.00)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.26)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.65)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded 1.09)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.99)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.87)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.16)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.16)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.16)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.16)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.46)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.97)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 154
\-------------------------

Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7364; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.30)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.12)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.40)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.40)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.40)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.40)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.40)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.90)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.32)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.32)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.32)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.32)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.32)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.32)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.13)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 155
\-------------------------

Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7349; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.58)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.87)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.51)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.87)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.68)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.81)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.81)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.81)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.81)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.98)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.92)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.92)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.58)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.58)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.58)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.58)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.42)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 156
\-------------------------

Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7334; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.77)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.11)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.81)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.58)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.55)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.26)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.89)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.62)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded 1.43)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.56)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.56)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.56)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.01)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.78)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.33)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.33)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.73)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.57)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.56)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 157
\-------------------------

Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7320; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.55)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove left instead of forward. (rewarded 0.29)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.15)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.68)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.85)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.18)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.06)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.84)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.31)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.09)
63% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.09)
63% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.09)
63% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.09)
63% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.09)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 0.95)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.94)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.15)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.27)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.27)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.62)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.62)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.58)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.32)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.32)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.06)
33% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.06)
33% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.06)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.36)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.09)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.06)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.32)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent drove forward instead of right. (rewarded -0.33)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent drove forward instead of right. (rewarded -0.33)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent drove forward instead of right. (rewarded -0.33)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.42)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.88)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.36)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.36)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.79)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 158
\-------------------------

Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7305; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.22)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove left instead of right. (rewarded 0.55)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.25)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.25)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.26)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.27)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.02)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.32)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.76)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.76)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.76)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.35)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.59)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.33)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.61)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.29)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.69)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.66)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.66)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.11)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.11)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.11)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.11)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.34)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.37)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.65)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.48)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.49)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.81)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.89)
7% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 159
\-------------------------

Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7291; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.15)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.09)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.65)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.65)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.65)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.83)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.83)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.83)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.83)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.89)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.75)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.34)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 1.27)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 1.27)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.78)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.72)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.50)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.49)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.30)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.30)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove forward instead of left. (rewarded 0.06)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.85)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.85)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.85)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.85)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.85)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.42)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.28)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.28)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.11)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.46)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.25)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 160
\-------------------------

Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7276; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.27)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.19)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.29)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.06)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.06)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.26)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.04)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.03)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.03)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.37)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.27)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.09)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.64)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.84)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.84)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded -0.26)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.01)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.29)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.09)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.09)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.09)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.09)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.80)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.68)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 161
\-------------------------

Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7261; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.58)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.69)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.80)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.61)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.70)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.41)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.75)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.58)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.15)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.89)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.86)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 162
\-------------------------

Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7247; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.63)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.63)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.76)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.12)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.74)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -11.00)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -11.00)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -11.00)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -11.00)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -11.00)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.07)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.33)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.85)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.22)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.23)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 1.47)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.99)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.61)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.22)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.22)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.22)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.22)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.75)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.27)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 163
\-------------------------

Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7233; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.73)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.39)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.36)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.65)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.20)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.20)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.84)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.04)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.79)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.71)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.71)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.71)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.71)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.71)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.82)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.36)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded -0.08)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.24)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.35)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.19)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.19)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.19)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.19)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.19)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.19)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.97)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.39)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 164
\-------------------------

Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7218; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.49)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.14)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.07)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 0.51)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.49)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.49)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.49)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.13)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.40)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.40)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.37)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.37)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.37)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.37)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.37)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.37)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.37)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.37)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.73)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.69)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.96)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.96)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.96)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.28)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.80)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.81)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.12)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded -0.34)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.10)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.46)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.46)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.71)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 165
\-------------------------

Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7204; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.83)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.50)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.13)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.21)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.75)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.45)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.47)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.38)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.34)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.60)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.61)
57% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 166
\-------------------------

Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7189; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.97)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.42)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.73)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.86)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.65)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.65)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.71)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.06)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.06)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.66)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.84)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.20)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.19)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.81)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.81)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.81)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.06)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.06)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.10)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.78)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.29)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 167
\-------------------------

Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7175; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.67)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.62)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.49)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.16)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.84)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.84)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.84)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.84)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.14)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.55)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.12)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.12)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.12)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.12)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.02)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.12)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 168
\-------------------------

Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7161; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.86)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.98)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.63)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.01)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.17)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.50)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.53)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove forward instead of left. (rewarded 0.38)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.31)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.22)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.46)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.46)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.46)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.46)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.46)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.68)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.13)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.81)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.81)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.81)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.85)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.70)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.70)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.70)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.31)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.45)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.81)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 2.38)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 1.43)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.21)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.99)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.04)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.66)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove left instead of right. (rewarded -0.55)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.82)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 169
\-------------------------

Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7146; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.36)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.62)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.95)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.95)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.95)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.95)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.01)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.98)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.47)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.47)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.47)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.70)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.31)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.31)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.31)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.31)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.63)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.86)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.62)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent drove forward instead of right. (rewarded -0.10)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent drove forward instead of right. (rewarded -0.10)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent drove forward instead of right. (rewarded -0.10)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent drove forward instead of right. (rewarded -0.10)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.93)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.69)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded 0.39)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 1.74)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.02)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.09)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.09)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.09)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.36)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.36)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.36)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.36)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded 0.65)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.26)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 170
\-------------------------

Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7132; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.16)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.20)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.93)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.93)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.93)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.03)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.06)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.08)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.17)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.71)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.59)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.59)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.59)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded 0.71)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded 0.71)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded 0.71)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.31)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.74)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.74)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.74)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.74)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.50)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.01)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent drove forward instead of right. (rewarded 1.48)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.73)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.20)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded 0.32)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.31)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.25)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.25)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.25)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.25)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.25)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.56)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.65)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.24)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 171
\-------------------------

Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7118; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.28)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.21)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.59)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 0.77)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.55)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.99)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.35)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.35)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.70)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.41)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.41)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.41)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.12)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.54)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.01)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.40)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.11)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.11)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.11)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 0.43)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 0.43)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 0.02)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.66)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.66)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.66)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.66)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.66)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.03)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 172
\-------------------------

Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7103; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.01)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.06)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.33)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.28)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.93)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.94)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent attempted driving left through a red light. (rewarded -9.53)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.81)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.27)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.27)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.37)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.78)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.98)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.61)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'right', None)
Agent drove forward instead of left. (rewarded -0.37)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.19)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.83)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 173
\-------------------------

Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7089; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.51)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.44)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.47)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove forward instead of right. (rewarded 0.30)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.49)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.34)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.34)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.34)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.41)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.49)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.31)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 174
\-------------------------

Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7075; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.75)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.96)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 0.44)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 0.44)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.25)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.58)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.64)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.78)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.58)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.77)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.68)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.20)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.20)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.20)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.20)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.11)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded -0.10)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.06)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.06)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.80)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.80)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.55)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded -0.84)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 175
\-------------------------

Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7061; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.59)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.54)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.46)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.20)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.10)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.09)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.70)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.70)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.56)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.62)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.74)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.03)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.42)
43% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 176
\-------------------------

Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7047; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.14)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.74)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.86)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.04)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.56)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.58)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.70)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.70)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.70)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.02)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.34)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.63)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.24)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 177
\-------------------------

Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7033; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.83)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.79)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.03)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.68)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.71)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.91)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.90)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.79)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.40)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.90)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.24)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.41)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.83)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'right')
Agent drove right instead of left. (rewarded -0.21)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.66)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.66)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.50)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.55)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.59)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.04)
4% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 178
\-------------------------

Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7019; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.27)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.01)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.70)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.79)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.14)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.98)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.97)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 179
\-------------------------

Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.7005; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded 1.64)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.58)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.97)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.74)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.74)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.88)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.09)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.28)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.47)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.30)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.91)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.91)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.10)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.02)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.43)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.95)
8% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 180
\-------------------------

Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6991; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.03)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.68)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.13)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.45)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.45)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.07)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.06)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'right')
Agent drove right instead of left. (rewarded 0.30)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.90)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.92)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.97)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.38)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.58)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.65)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'right', 'left')
Agent drove right instead of left. (rewarded 0.07)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.18)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.33)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 181
\-------------------------

Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6977; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.92)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.79)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.24)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.18)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.53)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.58)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.54)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.68)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.59)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.08)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 182
\-------------------------

Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6963; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.58)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.80)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.36)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.25)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.04)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.75)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.45)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.46)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.73)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.62)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded -0.03)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.58)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.36)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.41)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.24)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.24)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.24)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.14)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.14)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.14)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.53)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.80)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.17)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.56)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.36)
3% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 183
\-------------------------

Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6949; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.63)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.84)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.14)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.70)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.70)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.07)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.07)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.07)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.07)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.07)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.43)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.98)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.54)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.23)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.74)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.09)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.18)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 1.15)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.59)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.09)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.24)
5% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 184
\-------------------------

Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6935; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.01)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.82)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.82)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'right')
Agent drove forward instead of left. (rewarded 0.36)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove right instead of left. (rewarded 1.86)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.82)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.09)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.61)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.27)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.45)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.76)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.57)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 0.77)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.38)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.65)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.96)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.61)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.61)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.61)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.40)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 185
\-------------------------

Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6921; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.23)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.06)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.06)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.06)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.06)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.06)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.60)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.81)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.17)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.17)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.17)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.17)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.17)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.72)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.46)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.47)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.50)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.26)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.26)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.26)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.26)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.26)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.26)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.86)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.37)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 1.02)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.28)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.95)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.14)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 186
\-------------------------

Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6907; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.25)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.32)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.51)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.50)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.94)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.09)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.67)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.67)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.67)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.67)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.04)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.04)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.93)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.01)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.97)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.32)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.32)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.83)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.77)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.40)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.56)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.56)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.20)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.38)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.53)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.72)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 187
\-------------------------

Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6894; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.78)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.67)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.38)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 1.13)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.08)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.08)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.95)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.95)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.95)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.95)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.95)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.95)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.95)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.39)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.42)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.42)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.42)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.89)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 0.95)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.57)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.22)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.01)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove left instead of right. (rewarded 0.42)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.86)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.14)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 2.09)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.38)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.65)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.30)
16% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 188
\-------------------------

Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6880; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.07)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.29)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.05)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.24)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.22)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.44)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.08)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.97)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.98)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent followed the waypoint right. (rewarded 2.28)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.44)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.44)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.44)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.44)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.44)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.44)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.44)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.26)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.71)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.81)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.92)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.43)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.62)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.41)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.73)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.40)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.75)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 189
\-------------------------

Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6866; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.67)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'left')
Agent drove forward instead of right. (rewarded 1.25)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 2.50)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.16)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.81)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.51)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.51)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 1.05)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded 1.59)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.62)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.60)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.91)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.19)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 0.86)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 0.86)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 0.86)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 0.86)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 0.86)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.95)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.36)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.83)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 190
\-------------------------

Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6852; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.38)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.38)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.69)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.91)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.91)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.91)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.91)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.11)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.37)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent followed the waypoint forward. (rewarded 1.72)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.97)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.18)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.15)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.18)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.56)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.49)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 191
\-------------------------

Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6839; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.91)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.10)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.44)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.44)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.95)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.12)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.04)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.89)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.20)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.57)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.61)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.94)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.94)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.62)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.75)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.75)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.35)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.16)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 192
\-------------------------

Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6825; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 0.58)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.55)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.50)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.70)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.65)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.23)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.23)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.80)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.58)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 193
\-------------------------

Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6811; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.52)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 1.65)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.88)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.66)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.91)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.93)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.82)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.85)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.62)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.62)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.33)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.75)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.16)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.35)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.11)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.11)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.27)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.27)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.27)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.07)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 194
\-------------------------

Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6798; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.06)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.25)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.50)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.50)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.47)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.87)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.37)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 0.88)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.06)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.06)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 0.93)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.35)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.29)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.29)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.29)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.10)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.54)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.87)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.50)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.57)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.57)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.73)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.39)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.39)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.39)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.39)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.83)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 195
\-------------------------

Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6784; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.53)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.81)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove forward instead of right. (rewarded 0.59)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.23)
84% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.23)
84% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.23)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.73)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.88)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.60)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.97)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.85)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.27)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.46)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.59)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.66)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.16)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.04)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 0.32)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.59)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'right', 'right')
Agent followed the waypoint right. (rewarded 2.08)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.70)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.72)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.40)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.53)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.31)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.88)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 196
\-------------------------

Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6771; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.16)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove left instead of right. (rewarded 1.03)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.64)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.64)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.64)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.35)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.35)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.35)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.35)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded -0.05)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.83)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.21)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent drove left instead of forward. (rewarded 0.97)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.43)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.14)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.17)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.69)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.78)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.81)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.77)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.65)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.22)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.17)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 197
\-------------------------

Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6757; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 1.13)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.12)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.76)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.40)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.16)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.16)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.16)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.38)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.01)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.79)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.79)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.79)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.79)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.79)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.79)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.79)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.79)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.79)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.05)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.97)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.96)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.28)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.15)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.44)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.20)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.49)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.97)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.88)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 198
\-------------------------

Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6744; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.02)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.52)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.08)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.99)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.79)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.49)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.75)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.97)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.49)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.28)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.68)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.28)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.76)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.14)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded -0.12)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.24)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.75)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.41)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.50)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.38)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.80)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.24)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.33)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.54)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 199
\-------------------------

Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6730; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.35)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.04)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.28)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.28)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.76)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.76)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.76)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.83)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.25)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.25)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.88)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.41)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.87)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.46)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.37)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.41)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 0.03)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.93)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.54)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.36)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.43)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.43)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 200
\-------------------------

Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6717; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.05)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.75)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.60)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.58)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
86% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
86% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
86% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
86% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.69)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.22)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.22)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.78)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.78)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.78)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.78)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.50)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.42)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.39)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.41)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.93)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.93)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.48)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.48)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.48)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.70)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.70)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.26)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.26)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent drove right instead of left. (rewarded 0.38)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.01)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.87)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.86)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.37)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.37)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.37)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.37)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.29)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.99)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent drove right instead of left. (rewarded 1.07)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.36)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.04)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.14)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.44)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.60)
6% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.60)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.67)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.67)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.67)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.62)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 201
\-------------------------

Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6703; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.26)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.59)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.06)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.32)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.63)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.04)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.14)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.46)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.66)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.38)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.75)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.75)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.47)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove left instead of right. (rewarded 0.86)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.06)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.06)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.22)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 202
\-------------------------

Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6690; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 1.74)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.53)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.48)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.48)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.48)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.48)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.16)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.07)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.96)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.20)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.73)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.06)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.54)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.54)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent followed the waypoint forward. (rewarded 2.05)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent followed the waypoint forward. (rewarded 2.05)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent followed the waypoint forward. (rewarded 2.05)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent followed the waypoint forward. (rewarded 2.05)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.24)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.40)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.68)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.00)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.57)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.70)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.03)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.21)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.21)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.00)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 203
\-------------------------

Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6676; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.98)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.47)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.49)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.49)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.49)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.49)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.26)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.77)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.82)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.64)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.27)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.66)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.66)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.66)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.66)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.66)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.66)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.42)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.47)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.29)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.28)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.28)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.28)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.16)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.72)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 204
\-------------------------

Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6663; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove left instead of right. (rewarded 1.19)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.37)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.30)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.01)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.86)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 2.86)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.76)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.53)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.57)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.04)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.22)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.29)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 205
\-------------------------

Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6650; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.50)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.70)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.86)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 1.92)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.20)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.00)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.78)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.84)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.97)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.31)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.31)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.31)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.31)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.02)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.59)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.59)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.25)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 0.73)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.27)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.31)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.31)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.14)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.14)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.17)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.00)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.00)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.16)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.23)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.23)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.23)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 0.47)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.54)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.99)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.29)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 206
\-------------------------

Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6637; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.13)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.15)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.48)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.59)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.17)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.17)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.71)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.45)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.11)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.65)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.36)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.71)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.21)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded -0.17)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.49)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.77)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.77)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.77)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.77)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.77)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.07)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.75)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.12)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.59)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.06)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.33)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 207
\-------------------------

Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6623; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.10)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.90)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -11.00)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.17)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.49)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.49)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.58)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.44)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.44)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.99)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.73)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.33)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.34)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.56)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.56)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.56)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.60)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.02)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.02)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.02)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.48)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.48)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.48)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.61)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 1.61)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.68)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.32)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.38)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.38)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.38)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.52)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.08)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.37)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.60)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 208
\-------------------------

Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6610; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent drove forward instead of right. (rewarded 1.88)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove left instead of right. (rewarded 1.26)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.92)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.44)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.46)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 1.29)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 1.29)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 1.29)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.82)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.79)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.50)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.11)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 0.59)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.28)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.53)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.91)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.53)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.03)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.03)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.03)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.03)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.03)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.02)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.87)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.66)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.67)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.19)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.36)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.39)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 209
\-------------------------

Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6597; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.97)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.03)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.78)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.78)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.38)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.19)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.21)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 1.86)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 1.86)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded 0.96)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.08)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.97)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.97)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 0.91)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 0.91)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 0.91)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 0.91)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.39)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.43)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.63)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.17)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.59)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.59)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.59)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.59)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.59)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.56)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.27)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.00)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.83)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.94)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.94)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.94)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.94)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.67)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 210
\-------------------------

Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6584; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.46)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.35)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.81)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.25)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.53)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.61)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.55)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.85)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'right')
Agent followed the waypoint forward. (rewarded 2.51)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 211
\-------------------------

Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6570; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.08)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.85)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.71)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.30)
84% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.30)
84% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.30)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.44)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 1.52)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 1.52)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 1.52)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 1.52)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 2.00)
72% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 212
\-------------------------

Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6557; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.78)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.43)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.08)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.73)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.95)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.68)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.76)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.76)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.76)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.76)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.21)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.82)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.31)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.01)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.60)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.45)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.29)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.44)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.73)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.34)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded -0.35)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.44)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.44)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.44)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.44)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.44)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.44)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.85)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 213
\-------------------------

Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6544; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.20)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.19)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.66)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.66)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.66)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.66)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.66)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.66)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.66)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.32)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.84)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 214
\-------------------------

Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6531; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.18)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.01)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.01)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.01)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.01)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.16)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.16)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'right')
Agent attempted driving left through a red light. (rewarded -10.10)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.57)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.27)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.27)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.27)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.20)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.20)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.20)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.30)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 1.57)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.60)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.23)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.23)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.23)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.23)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.29)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.04)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.62)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.62)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.62)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.62)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.62)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.62)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.62)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.81)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.79)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.54)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 215
\-------------------------

Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6518; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.00)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.48)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.48)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.73)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.28)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.49)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.49)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.49)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.80)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.80)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.80)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.61)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.24)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 216
\-------------------------

Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6505; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.20)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.36)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.68)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.87)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.87)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.25)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.40)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 2.50)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.63)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.63)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.63)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.63)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.63)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.75)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.58)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.70)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.70)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.04)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.06)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.89)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.17)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.14)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.16)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.16)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.16)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.16)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.16)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.16)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.16)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.19)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.94)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.00)
16% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 217
\-------------------------

Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6492; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.86)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove left instead of right. (rewarded 0.80)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.02)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.71)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.73)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.13)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.44)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove left instead of right. (rewarded 1.32)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.65)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.27)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.12)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.88)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.66)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.70)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
33% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.18)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.18)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.33)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.01)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.07)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.41)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded -0.32)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.96)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.96)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.96)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded -0.32)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.91)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.06)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 218
\-------------------------

Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6479; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.32)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.30)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.11)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.59)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.83)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.83)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.83)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.83)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.81)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.13)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.64)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.74)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.13)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.68)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded -0.23)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.62)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.45)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.77)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.77)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.77)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.77)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.23)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.99)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.65)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 219
\-------------------------

Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6466; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.69)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.83)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.81)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.91)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.05)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.18)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.18)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.18)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.99)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.17)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 0.90)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.23)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.78)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.30)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.73)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.58)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.69)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.04)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.86)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 220
\-------------------------

Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6453; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 2.37)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.16)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.28)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.03)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.03)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.03)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.03)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.03)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.54)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.36)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.87)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.87)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.87)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.87)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.76)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 221
\-------------------------

Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6440; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.33)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.35)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.27)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.30)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.32)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.79)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.79)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.64)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 222
\-------------------------

Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6427; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent drove forward instead of right. (rewarded 1.17)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.12)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.12)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.12)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.12)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.08)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.08)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.89)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.87)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.87)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.87)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', 'right')
Agent drove right instead of forward. (rewarded 1.20)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.62)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.62)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.62)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.62)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.19)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.23)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 0.92)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.09)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.18)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.11)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.11)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.11)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.11)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.11)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.37)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.75)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 223
\-------------------------

Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6415; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.61)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.44)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.10)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent followed the waypoint forward. (rewarded 1.89)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.29)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.29)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.29)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.29)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.29)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.72)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.99)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.38)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent drove forward instead of left. (rewarded 1.49)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.41)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.67)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.90)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.85)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.85)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.01)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.01)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.90)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.90)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.13)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 224
\-------------------------

Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6402; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent followed the waypoint right. (rewarded 2.72)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.41)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.03)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 2.25)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.17)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.35)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.37)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.45)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.45)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.05)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.82)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.45)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.65)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.15)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.15)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.15)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.15)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
47% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
47% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
47% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent drove forward instead of left. (rewarded 0.74)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.71)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.71)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.71)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.15)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.15)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.17)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.17)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.17)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.17)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.69)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.29)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.79)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.34)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.87)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 0.27)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.59)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent drove right instead of left. (rewarded -0.05)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 225
\-------------------------

Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6389; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.61)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.63)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.01)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.01)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.01)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.01)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.25)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.02)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.04)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.88)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.88)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.88)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.31)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.18)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.65)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.30)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.36)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 226
\-------------------------

Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6376; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.11)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.55)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.55)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.55)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.55)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.55)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.63)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 1.45)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.65)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.86)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.53)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.53)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.53)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.43)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.50)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.17)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.36)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.51)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.51)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.69)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.46)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.46)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.46)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.46)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.46)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'right', 'left')
Agent followed the waypoint right. (rewarded 1.91)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.59)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.88)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.88)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.07)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.36)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded -0.45)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.06)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent properly idled at a red light. (rewarded -0.10)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 227
\-------------------------

Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6364; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.31)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.15)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.19)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.00)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.00)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.00)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.00)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.58)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.58)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.85)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.38)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.43)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.93)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.59)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.59)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.59)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.25)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.46)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.04)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.01)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.92)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.06)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.78)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.78)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.78)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.78)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.78)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.24)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.59)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 228
\-------------------------

Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6351; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.67)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.17)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.54)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.54)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.54)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.54)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.39)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.68)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.23)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.21)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.44)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.44)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.44)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.05)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.35)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.47)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.47)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.47)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.47)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.47)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.88)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.20)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.20)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.20)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.20)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.20)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.65)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 0.36)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.95)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.91)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.92)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 229
\-------------------------

Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6338; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.60)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.22)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.39)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.33)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.56)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.99)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.90)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.43)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.43)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.43)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.43)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.43)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.68)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.71)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.08)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.94)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.96)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.99)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 230
\-------------------------

Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6325; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.56)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.39)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.57)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.52)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.60)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.29)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.82)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.41)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.55)
47% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.55)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'right')
Agent drove left instead of forward. (rewarded 1.43)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.65)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.13)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.09)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.04)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.04)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.04)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.04)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.04)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.29)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.34)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.34)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.34)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.34)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.34)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove forward instead of left. (rewarded -0.05)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.53)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.38)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent drove forward instead of left. (rewarded 0.81)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.12)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.66)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.06)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 231
\-------------------------

Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6313; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.57)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent drove forward instead of right. (rewarded 1.84)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.74)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.74)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.74)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.74)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.63)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.79)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.73)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.70)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.37)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.89)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.89)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.46)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.46)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.46)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.60)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.43)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.43)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.43)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.43)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.65)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.15)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.15)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.62)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.69)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.18)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 0.04)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.26)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.57)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.57)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.57)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.68)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 0.56)
4% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 232
\-------------------------

Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6300; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.32)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.22)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.05)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.05)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.05)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.05)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.05)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.05)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 1.94)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.34)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.86)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.56)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.63)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.16)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.88)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.88)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.76)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.05)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.51)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.74)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.74)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.74)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.32)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.72)
51% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.72)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded 1.26)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.71)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.56)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.56)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.18)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.22)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.68)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.80)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 2.44)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.38)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.21)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.10)
11% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.10)
11% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.10)
11% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.10)
11% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.10)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.77)
6% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.77)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.19)
3% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.19)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 233
\-------------------------

Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6288; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.26)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.25)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.86)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.35)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.35)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.35)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.35)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.04)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.50)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.50)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', None)
Agent drove right instead of left. (rewarded 0.03)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.79)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.79)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.31)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.04)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.04)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.04)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.04)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.03)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.71)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.90)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.50)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.49)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.11)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.01)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.15)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 234
\-------------------------

Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6275; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.76)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove forward instead of left. (rewarded 0.76)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove forward instead of left. (rewarded 0.76)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.14)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.14)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.14)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.14)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.50)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.28)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.89)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.41)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.88)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.48)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.26)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.94)
43% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.94)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.57)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.18)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.39)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.39)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.39)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.39)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.39)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.72)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.16)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.90)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.90)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.90)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.90)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.62)
17% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 235
\-------------------------

Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6263; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.94)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.60)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.16)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded 1.53)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.96)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.33)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.38)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'right')
Agent followed the waypoint right. (rewarded 1.05)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.53)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.87)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.40)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 236
\-------------------------

Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6250; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.55)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.92)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded 0.52)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.46)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.09)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.09)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.09)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.09)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.59)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.03)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.31)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.23)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 237
\-------------------------

Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6238; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.17)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.91)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.15)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.15)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.15)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.15)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.15)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.75)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.28)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.28)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.98)
76% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 238
\-------------------------

Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6225; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.96)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.36)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.87)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.95)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.64)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent followed the waypoint forward. (rewarded 1.86)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.54)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 1.07)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.38)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.06)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.28)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.28)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.28)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.28)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.40)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.88)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.51)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.21)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.43)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 239
\-------------------------

Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6213; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.77)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.96)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.61)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.57)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.51)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.66)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.58)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.70)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.79)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.32)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.56)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.00)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.40)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent drove right instead of left. (rewarded 0.75)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.02)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 240
\-------------------------

Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6200; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.66)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.88)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.38)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.84)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.65)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.57)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.27)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.76)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.56)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.20)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.14)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.59)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.65)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.62)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.62)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.62)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.49)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.62)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.80)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.80)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.80)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.80)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.80)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.21)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.69)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.92)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.28)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded -0.11)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded -0.46)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 241
\-------------------------

Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6188; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.08)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.89)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.10)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.75)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.94)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.32)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.38)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.38)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.38)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.38)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.37)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.75)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.56)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.06)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.63)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.54)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.84)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.89)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.22)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.12)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.12)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.12)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.12)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.12)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.48)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove left instead of right. (rewarded -0.37)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded -0.61)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.58)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.42)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.60)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 242
\-------------------------

Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6175; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.87)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded 0.17)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.50)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.37)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.98)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.96)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.01)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.36)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.36)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.51)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.61)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.40)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.21)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.05)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.05)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.05)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.05)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.05)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.05)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.05)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.99)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.40)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.16)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 243
\-------------------------

Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6163; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.98)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.03)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.44)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.44)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.44)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.44)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.67)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.75)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.75)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.75)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.75)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.58)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.74)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.77)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.31)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.77)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.77)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.77)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.77)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.77)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.20)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.93)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -21.00)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.28)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent followed the waypoint left. (rewarded 1.41)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.58)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.18)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.18)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.52)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.07)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.88)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.55)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.55)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.55)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.85)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 0.72)
7% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 0.72)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.57)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.96)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 244
\-------------------------

Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6151; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.72)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.16)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.55)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.91)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.91)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.44)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.94)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.33)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded -0.04)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.50)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.50)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.50)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.50)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.46)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.98)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.38)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.06)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.83)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.06)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.02)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded -0.17)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.29)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 1.31)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 245
\-------------------------

Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6139; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.36)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.03)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.42)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.64)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.05)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.16)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.55)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 246
\-------------------------

Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6126; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 2.08)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.03)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.08)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.08)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.08)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.08)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.08)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.08)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.65)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.04)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.45)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.13)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.27)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.09)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.51)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.76)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 247
\-------------------------

Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6114; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.60)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.62)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.04)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.10)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.10)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.10)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.10)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.72)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.38)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.47)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 0.70)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.38)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.47)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.04)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.38)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 0.11)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.27)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.76)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.74)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 248
\-------------------------

Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6102; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.32)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.66)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.99)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.14)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 2.94)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.45)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.04)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.04)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.04)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.04)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.04)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.25)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.13)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.75)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.14)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.14)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.14)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.14)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.14)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.03)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.84)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove forward instead of left. (rewarded -0.23)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove forward instead of left. (rewarded -0.23)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.62)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.21)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.05)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.29)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.10)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.10)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.10)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.10)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.10)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.01)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.67)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.10)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.65)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.65)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.65)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.40)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.44)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 249
\-------------------------

Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6090; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.57)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 1.12)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.09)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.09)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.09)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.09)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.09)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.93)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.72)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.23)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.71)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.64)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.80)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.71)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.65)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.04)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.23)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.54)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.40)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove left instead of forward. (rewarded 0.04)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.70)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.70)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.70)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.99)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 250
\-------------------------

Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6077; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.90)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.06)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.75)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.66)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.71)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.71)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.36)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.96)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 251
\-------------------------

Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6065; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.47)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.60)
93% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.60)
93% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.60)
93% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.60)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.48)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.00)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.00)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.00)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.88)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.35)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.89)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.02)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.02)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.02)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.90)
67% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 252
\-------------------------

Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6053; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.54)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.28)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.14)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded -0.00)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.88)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.88)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.88)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.88)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.06)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.68)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.68)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.68)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.68)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.24)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 1.02)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 1.02)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 1.02)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 1.02)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 1.02)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.08)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.63)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.94)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.32)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.32)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.32)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.39)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.02)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.54)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.54)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.54)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.54)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.34)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.53)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.71)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.08)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.56)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
17% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
17% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
17% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
17% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.24)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.27)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 253
\-------------------------

Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6041; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.18)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.20)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.55)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.23)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.14)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.90)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.38)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.59)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.12)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.19)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.23)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.35)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.46)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.37)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.40)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.92)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.35)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.65)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.65)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.65)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.65)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.25)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 254
\-------------------------

Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6029; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.50)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.01)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.48)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.91)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 255
\-------------------------

Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6017; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.61)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.76)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.07)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.50)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.10)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.10)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.10)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.66)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.09)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.09)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.09)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.09)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.09)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.58)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.52)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove forward instead of left. (rewarded -0.12)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.76)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.30)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.59)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.90)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.28)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.38)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.38)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.38)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.51)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.29)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.58)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.18)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.81)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.81)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.81)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.60)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 0.81)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 256
\-------------------------

Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.6005; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.47)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.18)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.40)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.87)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.51)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.15)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.21)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.57)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.57)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.57)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.57)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.50)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.18)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.43)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.43)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.43)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.43)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.09)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.49)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.75)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.80)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.25)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.98)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.98)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.98)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.39)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.86)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.96)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.96)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.96)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.96)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.96)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.96)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.96)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.96)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.96)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.53)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 257
\-------------------------

Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5993; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.11)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.68)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.79)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.38)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.41)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 258
\-------------------------

Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5981; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.56)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 1.37)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.91)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.46)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.60)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.17)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.17)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.17)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.17)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.78)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded -0.19)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.36)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.36)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded -0.18)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded -0.18)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded -0.18)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded -0.18)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.79)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.34)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.56)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.82)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.08)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.63)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded -0.04)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.75)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 259
\-------------------------

Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5969; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.56)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.22)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.44)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.23)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.65)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.31)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.22)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.54)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove right instead of left. (rewarded 1.11)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.41)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.41)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.20)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.20)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent drove forward instead of right. (rewarded 0.48)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.31)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.57)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.24)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.93)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.93)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.93)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.93)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.54)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.40)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.72)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.50)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.99)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.99)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.30)
4% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 260
\-------------------------

Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5957; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.02)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.19)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.44)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.44)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.44)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.44)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.92)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.14)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.90)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.90)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.09)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.39)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.44)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.79)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.03)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.51)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.09)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.83)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 261
\-------------------------

Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5945; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.99)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.74)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.32)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.80)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 262
\-------------------------

Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5933; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.79)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.78)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.82)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.82)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.82)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.89)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.26)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.26)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.26)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.23)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.39)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.89)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.89)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.61)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.61)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.61)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.21)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.08)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.08)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.12)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.18)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded -0.17)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded -0.17)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.23)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.75)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.22)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.19)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.61)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 263
\-------------------------

Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5921; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.72)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.00)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.63)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.14)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 1.84)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.70)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.10)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.02)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.02)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.22)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.75)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.62)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.62)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.62)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.11)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.73)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.34)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.02)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.02)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.02)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.02)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.46)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.00)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.17)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.59)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 264
\-------------------------

Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5910; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.43)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded 0.18)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.22)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.82)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.89)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.77)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.77)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'left')
Agent drove forward instead of left. (rewarded -0.04)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 0.64)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 1.07)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.54)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.98)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.55)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 265
\-------------------------

Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5898; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.84)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.66)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.82)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.38)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.72)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.72)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.72)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.72)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.72)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.91)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.56)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.56)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.56)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.56)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.24)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.94)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.75)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.24)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.33)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.71)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.10)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 266
\-------------------------

Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5886; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.04)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.87)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.02)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.02)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.02)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.02)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.02)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.70)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.51)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.51)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.51)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove forward instead of left. (rewarded 1.17)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.53)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.53)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.53)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.20)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.26)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.81)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.10)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.20)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.62)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.72)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.72)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.50)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.59)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.34)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.45)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.22)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.58)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.58)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.58)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.58)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.58)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.76)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent drove right instead of left. (rewarded 1.28)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.29)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.29)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.29)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.29)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.29)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.29)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.01)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.09)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.96)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.25)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.56)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 267
\-------------------------

Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5874; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.20)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.49)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.44)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.44)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.44)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.44)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.44)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.44)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.73)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.63)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.71)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.71)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.40)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.44)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.01)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.20)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.20)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.39)
57% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 268
\-------------------------

Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5863; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.72)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.10)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.28)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.28)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.28)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.28)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.87)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.28)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.63)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.46)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 269
\-------------------------

Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5851; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.10)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.39)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.96)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.84)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.25)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.56)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.56)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.26)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.26)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.26)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.06)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'right')
Agent drove right instead of forward. (rewarded 0.34)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.13)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.64)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.83)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.60)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.28)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.62)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.57)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.30)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.87)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.44)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.43)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -9.58)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 270
\-------------------------

Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5839; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 1.28)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.32)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.08)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.87)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.76)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.73)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.99)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.99)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.99)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.90)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.51)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.49)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.48)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.19)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 271
\-------------------------

Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5827; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.92)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.84)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.01)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.51)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.83)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.50)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.88)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.88)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.88)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.88)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.88)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.30)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.49)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.66)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.04)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.10)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.10)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.98)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.98)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.98)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.98)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.98)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.98)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.50)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.29)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.34)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.30)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.24)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.97)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.20)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.32)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.44)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 272
\-------------------------

Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5816; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.61)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.41)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.93)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.71)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.05)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.18)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.60)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.40)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.40)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.40)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.40)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.63)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.17)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.17)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.17)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.17)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.17)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.17)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.16)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.17)
51% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.17)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.94)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.21)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.89)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.46)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.24)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.39)
29% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 273
\-------------------------

Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5804; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.40)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.01)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.80)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.97)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.97)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.97)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.97)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.01)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.94)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.94)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.94)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.94)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.38)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.82)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.85)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.16)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.05)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.35)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.79)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 274
\-------------------------

Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5793; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.91)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.19)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.67)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.19)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.62)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.02)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.05)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.72)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.19)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.31)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.01)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.18)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.86)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.88)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.88)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.88)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.88)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.88)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.64)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 275
\-------------------------

Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5781; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.43)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.15)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.69)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.25)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.47)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.48)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.50)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.14)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.88)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.82)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.49)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.49)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.49)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.49)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.60)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.47)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.09)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.77)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.23)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.87)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.27)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 276
\-------------------------

Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5769; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.26)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.44)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.51)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.87)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.84)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.48)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.41)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.45)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.42)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.40)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.40)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.40)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.40)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.41)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.04)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.96)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.96)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.96)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.96)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.96)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.03)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.28)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.28)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.28)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.98)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.11)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.10)
16% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 277
\-------------------------

Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5758; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.49)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.33)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.50)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.25)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.79)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.71)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.71)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.31)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.65)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.65)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent followed the waypoint right. (rewarded 1.68)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 0.84)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 278
\-------------------------

Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5746; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.19)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.76)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.88)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.88)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.88)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.00)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.70)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.90)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.02)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.56)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.56)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.30)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.30)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.23)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.15)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.15)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.15)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.82)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.04)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.04)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.09)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.87)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.58)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.11)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 279
\-------------------------

Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5735; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.34)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.73)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.20)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.20)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.98)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.98)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.98)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.98)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.98)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.66)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 1.14)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.41)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.80)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.13)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.11)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.11)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.29)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.67)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.13)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.13)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.13)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.13)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.41)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.33)
51% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.33)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.11)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
46% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.87)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.87)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.87)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.13)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 2.27)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 1.87)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.48)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.80)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.01)
26% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 280
\-------------------------

Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5724; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.58)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.11)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.79)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.18)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.60)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.71)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.71)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 281
\-------------------------

Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5712; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.85)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.17)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.56)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.72)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.75)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.33)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.75)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.80)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.42)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.34)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.34)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.34)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.34)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.39)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.39)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.51)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.51)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.03)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.84)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.84)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.84)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.84)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.75)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.11)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.80)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 282
\-------------------------

Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5701; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.87)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 0.56)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.79)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.77)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.96)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.26)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.40)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.72)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.92)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.92)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.92)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.92)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.92)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.84)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.21)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.74)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.49)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.49)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.70)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.70)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.70)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.70)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.70)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.70)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.84)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.47)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.46)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 283
\-------------------------

Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5689; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.94)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.04)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.66)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.97)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.80)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.39)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded 1.10)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.14)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.93)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.87)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.15)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.78)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.38)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.38)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.03)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.27)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.24)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 284
\-------------------------

Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5678; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.81)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.63)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.26)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.01)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.46)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.46)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.46)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.50)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.36)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.85)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.27)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.27)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.58)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.58)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.58)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.58)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.23)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.01)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.01)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.52)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.52)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.52)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.52)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.52)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.64)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.08)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.86)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.72)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.72)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 0.31)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.78)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 285
\-------------------------

Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5667; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.24)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.71)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.81)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 1.88)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 1.88)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 1.88)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.68)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.79)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.10)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.10)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.41)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.95)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.95)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.95)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.03)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.71)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.71)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.71)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.52)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.83)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.49)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.63)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.77)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.02)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.44)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.61)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 0.60)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.04)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.04)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.04)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.04)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.71)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded -0.47)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 286
\-------------------------

Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5655; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 1.37)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.48)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.90)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.84)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.83)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.00)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.00)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.00)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.00)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.23)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.70)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.70)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.15)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.15)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.15)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.15)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.15)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.71)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.71)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.71)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.71)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.49)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 287
\-------------------------

Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5644; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.18)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.73)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.49)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.26)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.38)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.30)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.27)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.27)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.27)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.02)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.08)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.08)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.08)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.24)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.71)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.71)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.71)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.27)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.48)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.39)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.39)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.39)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.39)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.87)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.21)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.21)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.21)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.21)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.80)
46% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.80)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.32)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.86)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.62)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.46)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.69)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.78)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.32)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.32)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.32)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.32)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.53)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.75)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.13)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.33)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.28)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 288
\-------------------------

Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5633; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'right')
Agent drove left instead of right. (rewarded 0.10)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.01)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.01)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.01)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.11)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.11)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.11)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.10)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.60)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.60)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.60)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.02)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.93)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.93)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.93)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.93)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.10)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.13)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.66)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.83)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove left instead of forward. (rewarded 1.32)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 1.49)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.30)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.35)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.33)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.64)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 289
\-------------------------

Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5621; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.39)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.55)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.55)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.55)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.55)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.55)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.82)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.64)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.64)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.64)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.64)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.00)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.79)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.79)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.55)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.27)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 290
\-------------------------

Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5610; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.78)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.78)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.28)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.28)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.28)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.28)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.28)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.28)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.63)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.61)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.61)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.61)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.61)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.58)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.70)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.83)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.83)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.19)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.19)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.19)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.19)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.19)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.14)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.77)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.77)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.77)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.77)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent drove right instead of left. (rewarded 1.31)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent drove right instead of left. (rewarded 1.31)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent drove right instead of left. (rewarded 1.31)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 0.71)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.94)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.76)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.67)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.06)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.59)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 291
\-------------------------

Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5599; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.85)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.53)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.12)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.30)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.07)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.13)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.13)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.33)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.45)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 0.89)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.07)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.07)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.07)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.07)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 1.23)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.89)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.11)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.30)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.05)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.84)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.85)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 292
\-------------------------

Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5588; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.43)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 2.63)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.49)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.49)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.49)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.49)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.49)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.49)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.59)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.10)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.51)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.23)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.69)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.88)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.21)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.03)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.30)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 0.92)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.26)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.26)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.26)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.82)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.44)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.91)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 293
\-------------------------

Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5577; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.20)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.02)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.76)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.76)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.33)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.13)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.13)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.13)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.98)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.04)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.04)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.04)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.80)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.74)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.16)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.99)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.99)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.99)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove right instead of left. (rewarded 1.38)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.02)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 0.86)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.02)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.02)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.02)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.02)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.23)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.04)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.04)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.04)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.04)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.04)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.53)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.94)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded -0.29)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.20)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.41)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.00)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 294
\-------------------------

Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5565; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.43)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.88)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.88)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.13)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.97)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.97)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.97)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.10)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.43)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.43)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.29)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove forward instead of right. (rewarded 1.68)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.68)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 295
\-------------------------

Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5554; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.45)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.99)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.42)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.80)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.87)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.87)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.87)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.87)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.14)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.67)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.39)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.07)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.67)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.74)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.74)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.43)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.67)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.19)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.33)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 0.64)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 0.64)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 0.64)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 0.64)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 0.64)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.83)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.97)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.84)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.82)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.82)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.82)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.82)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.63)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.67)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 296
\-------------------------

Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5543; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.15)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.05)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.15)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.37)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.63)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.91)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.91)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.68)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.74)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded -0.02)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.16)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.16)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.16)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.16)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.16)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.21)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.55)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.55)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.55)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.55)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.50)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.53)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.53)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.53)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.53)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.53)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.33)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.29)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 297
\-------------------------

Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5532; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded 0.79)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.97)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.97)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.97)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.97)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.97)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 1.39)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.93)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.31)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.97)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.11)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.11)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.11)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.11)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.85)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.04)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.66)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.40)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.40)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.40)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.60)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.37)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.20)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.15)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 298
\-------------------------

Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5521; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.36)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.59)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.75)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.93)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.63)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.99)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.30)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.02)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.09)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.22)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.61)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.04)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.82)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.14)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.49)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.91)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 299
\-------------------------

Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5510; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.31)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove left instead of forward. (rewarded 1.46)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.19)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.61)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.61)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.61)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.66)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.70)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.83)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.83)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.78)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.93)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.11)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.23)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.18)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 1.61)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.20)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.37)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.37)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.37)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.37)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 1.04)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.31)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.88)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.37)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.24)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.61)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.70)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.47)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.47)
29% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.47)
29% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.47)
29% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.47)
29% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.47)
29% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.47)
29% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.47)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded -0.01)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.01)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.87)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded -0.18)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.41)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
9% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
9% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
9% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
9% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.63)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.17)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.17)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.17)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.51)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 300
\-------------------------

Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5499; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.84)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.02)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.70)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.44)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.04)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.30)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.24)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.24)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.24)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.24)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.85)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 301
\-------------------------

Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5488; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.82)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 1.31)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.85)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.37)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.34)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.59)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.06)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 302
\-------------------------

Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5477; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.26)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.34)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded 0.48)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 1.13)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.08)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.24)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.51)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 1.77)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.30)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.05)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.83)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.83)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.83)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.83)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.83)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.12)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.46)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.87)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.87)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.69)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.53)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.53)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.53)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.53)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.74)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.64)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.30)
27% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.30)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.60)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.30)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.00)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.72)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 303
\-------------------------

Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5466; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.67)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.33)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.34)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.48)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.48)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.48)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.48)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.48)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.12)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.82)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.82)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.82)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.82)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.57)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.89)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.33)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.59)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded -0.08)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.50)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.87)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.10)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.41)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.41)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.41)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.38)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.02)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.81)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.81)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.81)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.09)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.65)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.65)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.65)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.03)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.13)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 304
\-------------------------

Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5455; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.23)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.42)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.40)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.17)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.82)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.17)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.33)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.94)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.80)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.14)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.68)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.10)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.59)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.66)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 305
\-------------------------

Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5444; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.98)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.79)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.06)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.10)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.92)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.62)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.62)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.62)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.86)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.78)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.39)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.96)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.90)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.27)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded -0.19)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.03)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.03)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.03)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.03)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 0.97)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 306
\-------------------------

Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5434; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.88)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.49)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.64)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.97)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.10)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.00)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.16)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.16)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.16)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.28)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.84)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.14)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.14)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.14)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.14)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.14)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.45)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.96)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.96)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.96)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.87)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.29)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent followed the waypoint forward. (rewarded 1.20)
5% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 307
\-------------------------

Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5423; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.13)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.11)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.06)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.01)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.87)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.26)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.03)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.10)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.10)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.49)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.23)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.88)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.84)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.30)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.46)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 308
\-------------------------

Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5412; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.87)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.93)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.65)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.65)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.94)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.59)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.66)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.66)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.01)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.66)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.66)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.66)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.66)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.66)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.66)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.81)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.58)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.58)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.58)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.58)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.58)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.07)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.11)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.90)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.02)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 309
\-------------------------

Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5401; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.56)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.87)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.94)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.94)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.94)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.94)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.94)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.20)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.42)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.31)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.64)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.64)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.62)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.62)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.62)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.95)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.95)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.97)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.44)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.44)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.63)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.63)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.63)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.42)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 1.49)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'right')
Agent drove right instead of forward. (rewarded 0.50)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.49)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 1.58)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded -0.25)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.82)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.83)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.24)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.24)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.24)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.24)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.04)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.37)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.90)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.03)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.27)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 310
\-------------------------

Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5390; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.26)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.37)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.76)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.15)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.15)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.29)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.28)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.28)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.28)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.28)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.08)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.84)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.86)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.22)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.92)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.59)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 311
\-------------------------

Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5379; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.92)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.74)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.20)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.11)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.39)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 312
\-------------------------

Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5369; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.37)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.36)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.85)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.85)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.85)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.85)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.85)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.95)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.18)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.52)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.12)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.81)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.47)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.22)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.66)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.77)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.78)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.68)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.96)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent drove forward instead of right. (rewarded 1.39)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.66)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.38)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.38)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.27)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.79)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.79)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.79)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.79)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.90)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.27)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 313
\-------------------------

Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5358; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.72)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.58)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.43)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.88)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.28)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.27)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.32)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.32)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.32)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.16)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.15)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.48)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.57)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.57)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.57)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.57)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.99)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.59)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.33)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.09)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 314
\-------------------------

Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5347; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.71)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.55)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.87)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.37)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.32)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.86)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.12)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.07)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.63)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.93)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.14)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.05)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.77)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.77)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.77)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.58)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.10)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.31)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.31)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 0.58)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 315
\-------------------------

Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5337; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.46)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.25)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.63)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.63)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.63)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.63)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.63)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.92)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.65)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.41)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.62)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.62)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.62)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.62)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.62)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.10)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.11)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.11)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.11)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.11)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.11)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.11)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.37)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.53)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.60)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.65)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.32)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.11)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.49)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 0.70)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 316
\-------------------------

Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5326; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent drove forward instead of right. (rewarded 1.56)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.90)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.40)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.56)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.02)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.51)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.51)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.95)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.82)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.82)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.82)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.82)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.64)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.75)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.52)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.13)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.64)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.71)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.13)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.58)
37% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.58)
37% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.58)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.71)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.85)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.08)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 1.23)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 1.23)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 1.23)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 1.23)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.65)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.07)
13% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.07)
13% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.07)
13% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.07)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.86)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.24)
7% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.24)
7% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.24)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 0.75)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 317
\-------------------------

Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5315; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.05)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.06)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.96)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.81)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.18)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.41)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.39)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.01)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.01)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.01)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.01)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.01)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.01)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.01)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.01)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.58)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 0.45)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.38)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.37)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.48)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.18)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 2.42)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.78)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.00)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.11)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.70)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 1.31)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 1.31)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.66)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 318
\-------------------------

Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5305; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.85)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.17)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.91)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.99)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.38)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.38)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.38)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.68)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.67)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.67)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded -0.06)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.45)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.23)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.23)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.23)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.23)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.23)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.23)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.35)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.40)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded -0.22)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.93)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.93)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.93)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.57)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.24)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.19)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.19)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.19)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.71)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.71)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.44)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.59)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.66)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.99)
8% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 319
\-------------------------

Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5294; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.28)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.62)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 1.83)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.63)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 0.21)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.20)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.37)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.37)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.47)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.47)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.84)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.01)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.43)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.43)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.43)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.85)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.85)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.85)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.85)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.54)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.36)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.36)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.36)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded -0.58)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.91)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 320
\-------------------------

Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5283; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.58)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.07)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.37)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.37)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.37)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.37)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.90)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.59)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.30)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.63)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.63)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.24)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.24)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.24)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.39)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.39)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.39)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.88)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.61)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.61)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 321
\-------------------------

Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5273; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.10)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.47)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.58)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.06)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.82)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 1.21)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 1.21)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 1.21)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 1.21)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.75)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.71)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.71)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.71)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.71)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 1.62)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.96)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.96)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.56)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.94)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.41)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.20)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.47)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.14)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.50)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.04)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.86)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.16)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent drove forward instead of left. (rewarded -0.11)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.32)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 322
\-------------------------

Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5262; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.84)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.95)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.17)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.66)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.66)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.66)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.66)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.66)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.88)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.84)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.56)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 0.44)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.39)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.33)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.78)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.78)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.78)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.78)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.65)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.98)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.47)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.20)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.20)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.57)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 1.01)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.61)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.81)
33% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.81)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.31)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.03)
27% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.03)
27% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.03)
27% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.03)
27% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.03)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.23)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.24)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 0.66)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.09)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.33)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.33)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.68)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.29)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 323
\-------------------------

Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5252; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.26)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.24)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.63)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.63)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.63)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.63)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.61)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.75)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.37)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.84)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.84)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.84)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.08)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
67% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 324
\-------------------------

Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5241; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 1.52)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.28)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.68)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.21)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.41)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.41)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.41)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.41)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.41)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.41)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.13)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.13)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.13)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 325
\-------------------------

Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5231; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.71)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.96)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.27)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.54)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.54)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.88)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.88)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.22)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.33)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove forward instead of right. (rewarded 1.70)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.48)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.28)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.93)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 326
\-------------------------

Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5220; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.00)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.27)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.16)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.79)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.98)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.51)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.60)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.60)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.60)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.26)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 327
\-------------------------

Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5210; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent drove right instead of forward. (rewarded 0.13)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.52)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.64)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.64)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.64)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.64)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.64)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.64)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.64)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.37)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.50)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.44)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.09)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.14)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.57)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.57)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.57)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.57)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.66)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.25)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.25)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.25)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 0.61)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.73)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded -0.48)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.38)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 328
\-------------------------

Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5200; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.85)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.31)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.08)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.08)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.84)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.19)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.19)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.43)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.32)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.06)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.53)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.36)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.18)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.43)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 0.89)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.98)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.75)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.15)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.84)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.40)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.88)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.88)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.88)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.88)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.53)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.00)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.05)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.09)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 329
\-------------------------

Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5189; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.74)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.99)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.60)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.56)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.56)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.56)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.46)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.12)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.07)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.22)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.98)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 0.99)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.80)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.80)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.80)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.10)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.60)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.60)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.60)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.60)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.94)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.43)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.29)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.19)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.68)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 330
\-------------------------

Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5179; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent attempted driving left through a red light. (rewarded -9.24)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.76)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.51)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.01)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.56)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.78)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.78)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.78)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.78)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.59)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.89)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.89)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.95)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.08)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.58)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.11)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded -0.06)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.48)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.48)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.48)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.48)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.87)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.41)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.68)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.31)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.23)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.10)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.53)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.64)
8% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 331
\-------------------------

Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5169; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.39)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.15)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.15)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.15)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.15)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.15)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.05)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.44)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'left')
Agent drove right instead of left. (rewarded 1.32)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.22)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.22)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.22)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.22)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.22)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.65)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.17)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.56)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.06)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.05)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.25)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.26)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.75)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.36)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 332
\-------------------------

Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5158; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.72)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.30)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.90)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.80)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.87)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.15)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.64)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.23)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.95)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.95)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.95)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.95)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.95)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.11)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.28)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.60)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.45)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.45)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.97)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.79)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.37)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.28)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.10)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.67)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.01)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 333
\-------------------------

Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5148; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.57)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.15)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.70)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.70)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.70)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.70)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.70)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.69)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.47)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.99)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.50)
72% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 334
\-------------------------

Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5138; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.96)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.24)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.35)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.60)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded 1.91)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded 1.91)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.22)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.50)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.50)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.50)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.50)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.18)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.67)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.29)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.74)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.05)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.05)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.56)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'right')
Agent followed the waypoint left. (rewarded 2.41)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', 'right')
Agent followed the waypoint left. (rewarded 2.41)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', 'right')
Agent followed the waypoint left. (rewarded 2.41)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', 'right')
Agent followed the waypoint left. (rewarded 2.41)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', 'right')
Agent followed the waypoint left. (rewarded 2.41)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 1.45)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.51)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.67)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.19)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.45)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 1.63)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.49)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.17)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.17)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.79)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.18)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.18)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.38)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.79)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.28)
7% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 335
\-------------------------

Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5127; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 1.19)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.14)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.90)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.08)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.84)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.72)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.27)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.33)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.77)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.77)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.77)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.77)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.36)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.39)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.39)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.39)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.21)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.04)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded 1.69)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded 1.69)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded 1.69)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.79)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.88)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.59)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent drove left instead of right. (rewarded 0.43)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove forward instead of right. (rewarded 0.13)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.53)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.53)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent drove forward instead of right. (rewarded -0.68)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.78)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.78)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.78)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.49)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.51)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 336
\-------------------------

Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5117; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.26)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent drove forward instead of left. (rewarded 0.43)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.14)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.14)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.14)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.12)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.33)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.33)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.05)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 0.91)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 0.91)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.82)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.01)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.01)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.28)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.30)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove forward instead of left. (rewarded 0.05)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove forward instead of left. (rewarded 0.05)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove forward instead of left. (rewarded 0.05)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove forward instead of left. (rewarded 0.05)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.10)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.55)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.98)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.02)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 0.08)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.62)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.23)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded -0.58)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.64)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.04)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.31)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 337
\-------------------------

Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5107; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.63)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.18)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.88)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.88)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.88)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.88)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.89)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 2.84)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.17)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.17)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.17)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.18)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.98)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 0.63)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 0.63)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 0.63)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 0.63)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.58)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.88)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.23)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 338
\-------------------------

Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5097; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.47)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.38)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.80)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.40)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.40)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.71)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.58)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.03)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.88)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.88)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.88)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.88)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.96)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.75)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.03)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.03)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.03)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.03)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.65)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.77)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.72)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.82)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.82)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.82)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.82)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.82)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.05)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.08)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.75)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 339
\-------------------------

Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5086; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.70)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.72)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.35)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.62)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.06)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.85)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.25)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.25)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.25)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.73)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.33)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.33)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.33)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.33)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.33)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.49)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.15)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 340
\-------------------------

Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5076; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.91)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.77)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.06)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.92)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.14)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.22)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.85)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.85)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.85)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.85)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.85)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.20)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.56)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.56)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.36)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.23)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.91)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.91)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.91)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.91)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.70)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.50)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 341
\-------------------------

Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5066; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.36)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.50)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.33)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.69)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'right', 'left')
Agent drove right instead of left. (rewarded -0.07)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.57)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.57)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.89)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.65)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.09)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.62)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.62)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.62)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.62)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.62)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.79)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.26)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.38)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.84)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.60)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.60)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.87)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.32)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.26)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.11)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded -0.48)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded -0.48)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded -0.48)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded -0.48)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded -0.48)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.04)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 342
\-------------------------

Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5056; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.13)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.00)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.24)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.24)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.24)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.24)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.86)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.86)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.23)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.23)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.23)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.46)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.18)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.80)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 343
\-------------------------

Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5046; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 1.27)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.17)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.16)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.26)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.58)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.09)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.09)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.39)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.42)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.22)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.22)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.22)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.22)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.22)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.71)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.13)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.13)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.03)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.71)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.36)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.03)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.97)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.96)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 344
\-------------------------

Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5036; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.42)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.77)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.71)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.78)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 0.34)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.07)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.07)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent drove forward instead of right. (rewarded 1.57)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.69)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.26)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.62)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.73)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.87)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.52)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.98)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.98)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.98)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.98)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 1.21)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.32)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.85)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove forward instead of right. (rewarded -0.24)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 345
\-------------------------

Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5026; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.52)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.92)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.06)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.00)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.88)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.88)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.88)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.88)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.57)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.74)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.96)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.43)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.62)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.41)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.41)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.62)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 346
\-------------------------

Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5016; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.64)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.45)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.63)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.89)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.24)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.40)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.40)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.40)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.40)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.13)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded 1.55)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.96)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.14)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.96)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.16)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.12)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.52)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.60)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.62)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.43)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.09)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.77)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.05)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded -0.53)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded -0.53)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded -0.53)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded -0.53)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.09)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 0.22)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 347
\-------------------------

Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.5006; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.65)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.65)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.65)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.65)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.65)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.49)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.37)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 1.14)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 1.14)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 1.14)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.22)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.44)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 348
\-------------------------

Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4996; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.42)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.01)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.89)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.44)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.44)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.52)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.85)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.85)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.68)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.68)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.51)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.21)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.21)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.21)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.57)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.14)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.10)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.40)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 349
\-------------------------

Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4986; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.13)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.78)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.19)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.48)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.91)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.73)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.13)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.55)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.55)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.55)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.55)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.77)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 350
\-------------------------

Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4976; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.77)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.81)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.49)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.64)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.40)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.66)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.53)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.49)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.00)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.33)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove left instead of right. (rewarded -0.17)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.08)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.34)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.34)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.34)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.34)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.77)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.65)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.56)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.16)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.19)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.99)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.11)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.47)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 351
\-------------------------

Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4966; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.85)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.50)
92% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.50)
92% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.50)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 1.97)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 1.97)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 1.97)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.58)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.48)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.04)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.66)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.64)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.98)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.79)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.52)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.80)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.18)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.29)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.28)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.28)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.28)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.12)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.81)
16% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 352
\-------------------------

Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4956; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.64)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.50)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.04)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.26)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 0.96)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.29)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.29)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.13)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.13)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 2.65)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 2.65)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 2.65)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.30)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.30)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.11)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.01)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.53)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.53)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.53)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.53)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.53)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.53)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.05)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.45)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded -0.56)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 353
\-------------------------

Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4946; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'right')
Agent followed the waypoint right. (rewarded 2.88)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.58)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.39)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.39)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.39)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.39)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.39)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove forward instead of left. (rewarded 0.12)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.65)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.32)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 0.92)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 354
\-------------------------

Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4936; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 0.32)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.43)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.44)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.44)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.44)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.44)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.44)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.44)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.65)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.90)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.64)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.64)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.64)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.64)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.92)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.54)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.58)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.93)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 355
\-------------------------

Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4926; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 0.21)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.72)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.55)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.01)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.11)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.73)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.58)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.55)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.66)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.85)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 356
\-------------------------

Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4916; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.89)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.20)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.26)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.58)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.27)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.19)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.16)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.07)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.18)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.76)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.50)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent drove right instead of forward. (rewarded 0.51)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.64)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.33)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.05)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 357
\-------------------------

Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4907; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.82)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.80)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.21)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.20)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.62)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.81)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.74)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.07)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.37)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.42)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.50)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.80)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.63)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.00)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.15)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.15)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.12)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.19)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.33)
5% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.33)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.43)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 358
\-------------------------

Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4897; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.34)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.70)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.43)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.05)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.60)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.60)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.60)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.70)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.18)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.18)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.18)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.18)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.18)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.36)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.69)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 359
\-------------------------

Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4887; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.80)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.24)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.61)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.61)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.00)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.11)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.63)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.19)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.10)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.58)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.58)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.58)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.58)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.58)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.58)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.80)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.11)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.11)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.63)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.12)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.11)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.14)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.79)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.32)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.30)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.10)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.06)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.57)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.27)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'right')
Agent followed the waypoint right. (rewarded 1.70)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.24)
7% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 360
\-------------------------

Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4877; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.77)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove left instead of forward. (rewarded 1.56)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.38)
91% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.38)
91% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.38)
91% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.38)
91% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.38)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.52)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.45)
86% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.45)
86% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.45)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.58)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.56)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.41)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.41)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.66)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.66)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.66)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.66)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.78)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.50)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.30)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.92)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 1.28)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.68)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.88)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove right instead of left. (rewarded -0.34)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove forward instead of right. (rewarded -0.15)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.93)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.72)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.50)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.50)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.50)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.57)
17% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 361
\-------------------------

Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4868; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.57)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.77)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent drove left instead of forward. (rewarded 0.33)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.46)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.20)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.24)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.24)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.24)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.77)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.79)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.72)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.29)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.85)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.55)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent drove right instead of forward. (rewarded -0.01)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.34)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.94)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.25)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.13)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.86)
5% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 362
\-------------------------

Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4858; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.32)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.00)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.87)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.50)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.32)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.54)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.94)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 363
\-------------------------

Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4848; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.90)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.90)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.90)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.90)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.27)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.45)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.77)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.75)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.31)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.10)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.72)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.34)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.34)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.51)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.12)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.13)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.40)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 364
\-------------------------

Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4838; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.68)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.29)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.60)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.20)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.18)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.18)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.11)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.83)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.03)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.32)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.29)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.30)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.51)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.11)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.28)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.28)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.87)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.34)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded -0.31)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'right')
Agent drove right instead of left. (rewarded -0.19)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.13)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.01)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 365
\-------------------------

Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4829; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.44)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.21)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.17)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.79)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.03)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.39)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.39)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.39)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.18)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.51)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.06)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.83)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.37)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.79)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.79)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.79)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.15)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.95)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 366
\-------------------------

Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4819; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.10)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.21)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.66)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.66)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.66)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.66)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.66)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.31)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove forward instead of right. (rewarded 1.06)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.90)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.90)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.90)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.54)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 367
\-------------------------

Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4809; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.68)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.06)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.41)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.73)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.73)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.71)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.71)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.71)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.71)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.90)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.41)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.47)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.92)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.48)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.12)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.95)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.04)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.63)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.78)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.61)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.13)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.13)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.13)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.13)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.31)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 368
\-------------------------

Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4800; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.11)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.50)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent drove right instead of left. (rewarded 1.05)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.67)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.61)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.83)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.89)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.91)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.78)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.57)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.79)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded -0.19)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.80)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.80)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.80)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.42)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.42)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.21)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.21)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.21)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.43)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.44)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.72)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.00)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 369
\-------------------------

Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4790; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.10)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.71)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.70)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.99)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.86)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.86)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.86)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.86)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.88)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.61)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.61)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.61)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.98)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.24)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 0.09)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.07)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.07)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.07)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.07)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.49)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 370
\-------------------------

Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4781; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.50)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.80)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.54)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.54)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.54)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.65)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.65)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.22)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.74)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.70)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.04)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.74)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.74)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.49)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.08)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 371
\-------------------------

Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4771; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.95)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.59)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.08)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.12)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.12)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.12)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.12)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.68)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.61)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 372
\-------------------------

Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4762; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.76)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.91)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.87)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.45)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.27)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.38)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.38)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.02)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.64)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.44)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.14)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.53)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.45)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.20)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.00)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'right')
Agent attempted driving left through a red light. (rewarded -9.07)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.38)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.25)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.66)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.48)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.48)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.48)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 0.61)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 0.61)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 0.61)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 373
\-------------------------

Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4752; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.65)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.06)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 2.69)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.90)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.92)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.77)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.10)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.10)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.10)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.10)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.83)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.79)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.23)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 374
\-------------------------

Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4743; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent drove forward instead of left. (rewarded 0.63)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.76)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.13)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.28)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.57)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.57)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.47)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.47)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.30)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.67)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.05)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.14)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.14)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.14)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.14)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.25)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.38)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.18)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.18)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.63)
37% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 375
\-------------------------

Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4733; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.35)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.61)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.16)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.99)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded -0.01)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.92)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.83)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.49)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.33)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.33)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.33)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.33)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent drove forward instead of right. (rewarded -0.19)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.36)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.36)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.36)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.36)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove forward instead of right. (rewarded 1.06)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.20)
36% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 376
\-------------------------

Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4724; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.41)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.56)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.44)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.53)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.28)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.33)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.60)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.43)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.07)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.09)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.09)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.24)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.28)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.15)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.38)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.38)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.38)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.38)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.50)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.09)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 377
\-------------------------

Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4714; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.04)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.69)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.12)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.12)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.12)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.42)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.42)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.42)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.42)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.99)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.87)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.33)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.33)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.33)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.33)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.33)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.37)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.37)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.74)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.74)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.74)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.74)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.99)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded -0.15)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.58)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.58)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.69)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.88)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.03)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.12)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.43)
5% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 378
\-------------------------

Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4705; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.50)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.44)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.23)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.23)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.45)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.99)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.70)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.70)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.70)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.70)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.74)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.60)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 379
\-------------------------

Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4695; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.21)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.98)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.98)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.98)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.98)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.98)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.76)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.00)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.00)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.00)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.00)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.89)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.56)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.81)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.81)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.81)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.81)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.81)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.51)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.76)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.67)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.89)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.54)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.32)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.11)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.71)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.66)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.66)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.61)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 380
\-------------------------

Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4686; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.65)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.76)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.41)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.74)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.25)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.23)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.95)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.87)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.79)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.14)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.98)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.20)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 381
\-------------------------

Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4677; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.22)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.81)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.49)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.70)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.70)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.70)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.70)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.17)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.14)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.14)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.04)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded 1.45)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.84)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.87)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded -0.17)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.27)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.68)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.68)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.68)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.68)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.35)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 0.85)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.20)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.20)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.20)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.43)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.16)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.16)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.16)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.16)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.16)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.10)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded -0.57)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.10)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent drove right instead of forward. (rewarded -0.13)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.76)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 382
\-------------------------

Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4667; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.21)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.27)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 1.07)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 1.07)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.35)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.40)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.86)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.49)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove forward instead of right. (rewarded 1.53)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.14)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 383
\-------------------------

Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4658; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.21)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.95)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.06)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.06)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.06)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.81)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.49)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.66)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.38)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 1.75)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 1.75)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 1.75)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.55)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.91)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'right')
Agent drove left instead of right. (rewarded 1.61)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.20)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.46)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.46)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove forward instead of right. (rewarded 0.12)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.74)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 384
\-------------------------

Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4649; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.25)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.59)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.32)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.16)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.16)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.77)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.37)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.37)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.04)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.89)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.89)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.89)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.62)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 385
\-------------------------

Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4639; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.78)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.62)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.13)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.13)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.13)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.92)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.30)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.30)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.18)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.18)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.18)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.18)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.77)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.75)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.27)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.18)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.42)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.42)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.42)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.42)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.15)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.15)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.46)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.46)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.83)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.19)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.01)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.06)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.61)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.75)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.64)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.64)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.01)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.01)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.01)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.38)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.18)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.11)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.04)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.24)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.37)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.72)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.04)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 386
\-------------------------

Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4630; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.66)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.69)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.60)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.58)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.62)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'right')
Agent drove right instead of forward. (rewarded 1.44)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.53)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.23)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.15)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.24)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.12)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.12)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.12)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.12)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.98)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.41)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.41)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.41)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.41)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.58)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.46)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded -0.26)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.36)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 387
\-------------------------

Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4621; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.01)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.33)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.10)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.79)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.38)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.22)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.22)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.22)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.17)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.73)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.73)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.51)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.51)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.10)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.10)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.10)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.80)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.49)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.90)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 1.34)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.15)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.17)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.29)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.71)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 388
\-------------------------

Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4612; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.29)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.15)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.61)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.22)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.22)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.43)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.12)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.41)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.25)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.61)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.61)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.61)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.71)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.71)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.71)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.15)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.15)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.54)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.16)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 0.94)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.36)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 389
\-------------------------

Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4602; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.49)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.18)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.24)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.28)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.28)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.28)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.28)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.28)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.73)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.54)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.54)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.93)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.19)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.91)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 0.17)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 0.17)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 0.17)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.47)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.11)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.86)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.77)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.65)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.65)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.06)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.07)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.92)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.82)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.82)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.62)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.68)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 390
\-------------------------

Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4593; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.82)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 1.83)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.66)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.36)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.56)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.35)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.44)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.51)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.96)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.73)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.10)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 1.12)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 2.67)
43% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 391
\-------------------------

Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4584; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.70)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.09)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.02)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.84)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.41)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 0.83)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 0.83)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.61)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.48)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.61)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.00)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.09)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.09)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.09)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.12)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.42)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.70)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.40)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.10)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.90)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 392
\-------------------------

Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4575; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent drove forward instead of right. (rewarded 1.15)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.23)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.07)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.17)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.17)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.57)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.57)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.90)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.12)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.69)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.76)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.61)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.59)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.71)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.70)
36% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 393
\-------------------------

Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4566; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.35)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.78)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.45)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.40)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.40)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.91)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.71)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.71)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.71)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.71)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.71)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.02)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.92)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.58)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.58)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.93)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.61)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.61)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.42)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.42)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.42)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.31)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.04)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.04)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.48)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.48)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.31)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.16)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 394
\-------------------------

Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4557; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.35)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.75)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.39)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.16)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 395
\-------------------------

Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4548; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.32)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.01)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.85)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.25)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.30)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.63)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.66)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.66)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.66)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.66)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.66)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.37)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 396
\-------------------------

Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4538; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.06)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.48)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.07)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.07)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.74)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.87)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.37)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.55)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.83)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.01)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.52)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 1.57)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.80)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.80)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.80)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.80)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.16)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.23)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.23)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.17)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.17)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.17)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.17)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.17)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.91)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.59)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded -0.46)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 397
\-------------------------

Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4529; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.51)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.84)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.00)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.13)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.92)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.24)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.24)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.24)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.84)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.78)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.15)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 0.83)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.31)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.31)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.31)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.31)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.31)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.31)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.72)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.08)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.29)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.35)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.84)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.84)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.84)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.84)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.77)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.76)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.50)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.62)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.28)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.54)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 398
\-------------------------

Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4520; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.73)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.27)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.37)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.37)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.37)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.37)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.37)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.12)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.11)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.61)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.61)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.61)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.61)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.61)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.82)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.08)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.59)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.74)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.73)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.11)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 399
\-------------------------

Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4511; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent drove forward instead of right. (rewarded 1.32)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.22)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
90% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
90% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
90% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.48)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.21)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.21)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.21)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.14)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.14)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.29)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.29)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.29)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.71)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.69)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.15)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.25)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.17)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.30)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.41)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.36)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.15)
43% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 400
\-------------------------

Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4502; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.80)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.18)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.57)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.15)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.15)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.47)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.88)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.03)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.17)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.55)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 401
\-------------------------

Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4493; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.75)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove left instead of right. (rewarded 1.10)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
90% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.98)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.45)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.06)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.06)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.06)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.06)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.06)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.06)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.94)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.18)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.50)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 1.04)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 1.04)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 1.04)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 1.04)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.70)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.13)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.99)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 402
\-------------------------

Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4484; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.44)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.88)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.42)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.36)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.36)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.36)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.36)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.36)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.63)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.59)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.47)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.33)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.29)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.78)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.35)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 403
\-------------------------

Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4475; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.89)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.67)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.73)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.70)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.76)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.09)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.15)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.78)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.41)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.62)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.67)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.00)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.73)
47% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 404
\-------------------------

Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4466; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.12)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.95)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.00)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.00)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.00)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.00)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.67)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.91)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.91)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.91)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.91)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.21)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove forward instead of left. (rewarded 0.60)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.48)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.70)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.91)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.44)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.02)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.49)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 2.04)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.54)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.05)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.07)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.58)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.22)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.22)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.22)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.59)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.42)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.25)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 405
\-------------------------

Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4457; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.51)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.13)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.33)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.33)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.91)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.49)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.49)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.49)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.49)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.35)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.30)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.47)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.43)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.43)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.43)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.43)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.49)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 1.53)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.85)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.75)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.57)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.62)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.62)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.62)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.62)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.70)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 0.82)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.84)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.09)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.08)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.25)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.65)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 406
\-------------------------

Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4449; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.58)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.07)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.72)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.88)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.59)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.68)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.72)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 1.64)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.46)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.70)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.35)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.06)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.01)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.02)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.64)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.64)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.64)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.64)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.64)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.62)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 407
\-------------------------

Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4440; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.09)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.16)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.20)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.52)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 408
\-------------------------

Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4431; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.15)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.65)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.51)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.51)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.41)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent followed the waypoint left. (rewarded 2.65)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.02)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.96)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.03)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.21)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.49)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.10)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.34)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.34)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.10)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.10)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.13)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.44)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.93)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.59)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 409
\-------------------------

Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4422; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.02)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.51)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 0.45)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.08)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.40)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.44)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.01)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.09)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.71)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.07)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.57)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.24)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.84)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.84)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.00)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.00)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.00)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.00)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.34)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.34)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.00)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.00)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.78)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.48)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.73)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 410
\-------------------------

Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4413; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.55)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.81)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.91)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.43)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.39)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.04)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent followed the waypoint forward. (rewarded 2.21)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.02)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'left')
Agent drove forward instead of left. (rewarded 0.22)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.58)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.48)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.69)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.74)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.36)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 411
\-------------------------

Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4404; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.01)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.46)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.69)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.69)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.98)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.41)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.06)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.65)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.65)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.95)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.25)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.25)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.19)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.19)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.19)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.50)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.72)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.41)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.41)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.41)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.41)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.91)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.07)
49% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 412
\-------------------------

Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4396; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.74)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.19)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.32)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.60)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 413
\-------------------------

Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4387; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.78)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 1.75)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.80)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.02)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.02)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.02)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.59)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.06)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.06)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.06)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.75)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.72)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent drove forward instead of right. (rewarded 1.53)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove forward instead of right. (rewarded 1.53)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove forward instead of right. (rewarded 1.53)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove forward instead of right. (rewarded 1.53)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.43)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 414
\-------------------------

Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4378; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.75)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.51)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.77)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.53)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.51)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.19)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.91)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.91)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.08)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.61)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.57)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.87)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.00)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.51)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.26)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.43)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.89)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 0.65)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 415
\-------------------------

Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4369; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.91)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.43)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.48)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent drove right instead of forward. (rewarded 1.02)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.76)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.30)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove left instead of forward. (rewarded 0.06)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.66)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.35)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.35)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.90)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.90)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.99)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.36)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.58)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.17)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 416
\-------------------------

Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4360; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.34)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.01)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.11)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.11)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.11)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.11)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.11)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.11)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 0.71)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.18)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.29)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.43)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.16)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.52)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.34)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.36)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.02)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove forward instead of left. (rewarded 1.28)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove forward instead of left. (rewarded 1.28)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded -0.08)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.13)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'right')
Agent followed the waypoint right. (rewarded 0.46)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.57)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'right')
Agent drove left instead of forward. (rewarded 0.86)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.60)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 417
\-------------------------

Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4352; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.94)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.33)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.47)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.48)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.56)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.37)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 418
\-------------------------

Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4343; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 1.64)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.03)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.16)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.41)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.22)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.98)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.14)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.48)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.22)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.75)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.10)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.42)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.13)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.11)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.40)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.46)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.69)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.35)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.44)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 419
\-------------------------

Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4334; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.38)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.07)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.07)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.07)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.54)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.78)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.38)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.89)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.89)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded -0.11)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.36)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.36)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.36)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.36)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.68)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 420
\-------------------------

Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4326; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.31)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.32)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.32)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.08)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.51)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.02)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.85)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.92)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 0.92)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 421
\-------------------------

Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4317; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.85)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.04)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.39)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.21)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.21)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.86)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.36)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 422
\-------------------------

Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4308; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 2.65)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.16)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.05)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.06)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.73)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.00)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.90)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 423
\-------------------------

Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4300; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 0.99)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.75)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.42)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.07)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.86)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.50)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.63)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent drove right instead of forward. (rewarded 1.17)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent drove right instead of forward. (rewarded 1.17)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.96)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 424
\-------------------------

Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4291; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.43)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.05)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.78)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.10)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.54)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.97)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.93)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 1.37)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 1.37)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 1.37)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 1.37)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.41)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.82)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.53)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.67)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.73)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.85)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.13)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.31)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.31)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.31)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.31)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.31)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.56)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.46)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.31)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.15)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.82)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.67)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.15)
7% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.15)
7% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.15)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.10)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 425
\-------------------------

Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4283; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.88)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.64)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.75)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.02)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.62)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.62)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.91)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.60)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.04)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.87)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.34)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.41)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 426
\-------------------------

Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4274; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.07)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.17)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.58)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.41)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.36)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.34)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.35)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.99)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.99)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.99)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.99)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.34)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.73)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 0.36)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.47)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.94)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.94)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.94)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.94)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.94)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.02)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.66)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.65)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.12)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 427
\-------------------------

Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4266; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.90)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.85)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.99)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.99)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.99)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.99)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.59)
74% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.59)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
71% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
71% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.88)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.20)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent drove right instead of left. (rewarded 1.29)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.53)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.72)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.11)
51% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.11)
51% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.11)
51% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.11)
51% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.11)
51% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.11)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.13)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.95)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.95)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.95)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.95)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.22)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.10)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 1.82)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.48)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.87)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.45)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.23)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.77)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.69)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.89)
14% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 428
\-------------------------

Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4257; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.85)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.68)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.29)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.32)
84% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.32)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.00)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.69)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent followed the waypoint left. (rewarded 2.61)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent followed the waypoint left. (rewarded 2.61)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.50)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.50)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.50)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 0.21)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.14)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.57)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.57)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.57)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.20)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 0.04)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.06)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.61)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded -0.12)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded -0.12)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.92)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.92)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.92)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.92)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.48)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.16)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.92)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.97)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 429
\-------------------------

Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4249; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.56)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.90)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.63)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.63)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.63)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.63)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.54)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.07)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.57)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.36)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.32)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.37)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.82)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.59)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.86)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.64)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.14)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.14)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.14)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.14)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.14)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.27)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.55)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.55)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.37)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.14)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.72)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.84)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.81)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.84)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.84)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.84)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.90)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 430
\-------------------------

Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4240; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.47)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.91)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.01)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.55)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.69)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.07)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded -0.02)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.11)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.90)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 431
\-------------------------

Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4232; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.34)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.57)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.51)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.97)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.68)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.44)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.86)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.76)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.76)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.30)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.63)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.31)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.56)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.10)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded -0.45)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.70)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.56)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.74)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 0.43)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 432
\-------------------------

Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4223; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.84)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.99)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.14)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.63)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.88)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.87)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.19)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.60)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.28)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.17)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.17)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.17)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.87)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.13)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.13)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.13)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.49)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.76)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.76)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.76)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.04)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.04)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.21)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.05)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.12)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.35)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.34)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded 1.10)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.88)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.68)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 433
\-------------------------

Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4215; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.84)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.76)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.69)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.73)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.65)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.53)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.53)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.53)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.57)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.62)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.62)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.62)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.13)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 0.97)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 434
\-------------------------

Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4206; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.01)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.46)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.52)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.48)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.50)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.50)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.50)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.50)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.57)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.68)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.31)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.14)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.69)
63% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 435
\-------------------------

Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4198; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.96)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.67)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.17)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.17)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.17)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.99)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.34)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.11)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.33)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.88)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.88)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.04)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.58)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 436
\-------------------------

Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4190; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.27)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.43)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.79)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.84)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.84)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.11)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.83)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.83)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 437
\-------------------------

Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4181; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'right')
Agent properly idled at a red light. (rewarded 2.80)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.51)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.26)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.91)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.71)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.57)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.79)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.79)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.79)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.79)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.79)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.79)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.82)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.59)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.59)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.59)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.59)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.81)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.31)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.47)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.23)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 0.73)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 0.73)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.64)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded -0.18)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded -0.18)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded -0.18)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.35)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.03)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.03)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.03)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.03)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.87)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.87)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.09)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.09)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.09)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.09)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.91)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 438
\-------------------------

Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4173; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.98)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.32)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.53)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.48)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.49)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.98)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.88)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.82)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.29)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.55)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 439
\-------------------------

Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4164; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'left')
Agent followed the waypoint left. (rewarded 2.09)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.47)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.80)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.72)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 440
\-------------------------

Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4156; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.36)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.53)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.49)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.35)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.12)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove left instead of forward. (rewarded 1.85)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove left instead of forward. (rewarded 1.85)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove left instead of forward. (rewarded 1.85)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove left instead of forward. (rewarded 1.85)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove left instead of forward. (rewarded 1.85)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.89)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.42)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.90)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.95)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.65)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 441
\-------------------------

Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4148; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.67)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 0.03)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.44)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.44)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.44)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.11)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.27)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.09)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.79)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.79)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.79)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.79)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.83)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.28)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.28)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.28)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.97)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.65)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.65)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.65)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.65)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.65)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.65)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.27)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.05)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.76)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.44)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.17)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.72)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 0.79)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 442
\-------------------------

Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4140; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.56)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.21)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.78)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.50)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.50)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.50)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.50)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.50)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.92)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.66)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.12)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.20)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.95)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.74)
69% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 443
\-------------------------

Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4131; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.23)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.07)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.70)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.70)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.70)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.98)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.53)
83% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.53)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.15)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.52)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.52)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.54)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.57)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.84)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.90)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.99)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.55)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.50)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.91)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.37)
43% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.37)
43% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.37)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.71)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 444
\-------------------------

Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4123; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.41)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.39)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.74)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.74)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.74)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.74)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.74)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.74)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.75)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.72)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.72)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.34)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.03)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.21)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.65)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.37)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.37)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.37)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.60)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.91)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.82)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 0.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 0.78)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.97)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.35)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.35)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.35)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.35)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.35)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.35)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.35)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.35)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.54)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.35)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.75)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.35)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 445
\-------------------------

Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4115; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 1.53)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.78)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.58)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.09)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.09)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.09)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.09)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.91)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.44)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.51)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.41)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.54)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.54)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.96)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.00)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.94)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.94)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.94)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.56)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.28)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.04)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 446
\-------------------------

Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4107; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.94)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.31)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.91)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.91)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.91)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.91)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.91)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.91)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.67)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.13)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.81)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.93)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 447
\-------------------------

Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4098; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.67)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.85)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.45)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.09)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.30)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.49)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.49)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.49)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.76)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.92)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.92)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.10)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.75)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.75)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.26)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.26)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.12)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.12)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.12)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 0.55)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.47)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.48)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.50)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.55)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.41)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.98)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.87)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.02)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.59)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.59)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.59)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.59)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.78)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.40)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 448
\-------------------------

Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4090; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.92)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.92)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.02)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.91)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.68)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.95)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.95)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.95)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.95)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.04)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.14)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.70)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.88)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.58)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.56)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.53)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.38)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.38)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.93)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.08)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.08)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.08)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 449
\-------------------------

Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4082; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded 1.38)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.55)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.04)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.79)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.23)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.40)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.73)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.72)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.68)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.68)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.02)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.33)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.33)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.33)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.33)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.33)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.60)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.17)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 450
\-------------------------

Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4074; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.07)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.79)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.05)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.05)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.05)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.12)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.68)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.36)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove left instead of forward. (rewarded 1.07)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.67)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.96)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.96)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.96)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.96)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.96)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.98)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.01)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.72)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.10)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.63)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.63)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.17)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.77)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 451
\-------------------------

Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4066; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.89)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.14)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.42)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.42)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.42)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.42)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.42)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.42)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.11)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.20)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.41)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.41)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.65)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 0.81)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.33)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.18)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.18)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.51)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.51)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.51)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.51)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.05)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.53)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.35)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -9.96)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 452
\-------------------------

Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4058; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.71)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.48)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.82)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.61)
89% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.61)
89% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.61)
89% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.61)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.89)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.19)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.18)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.59)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.59)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.59)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 1.67)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
71% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
71% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
71% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.87)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.94)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.28)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.11)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.46)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.05)
51% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.05)
51% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.05)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.75)
49% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 453
\-------------------------

Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4049; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.69)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.14)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.77)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.05)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.70)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.29)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.39)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 454
\-------------------------

Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4041; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.93)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.66)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.07)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.72)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.72)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.72)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.72)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.72)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.37)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.08)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.95)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.79)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.79)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.79)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.79)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.79)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.43)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.20)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.20)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.20)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.20)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.20)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.23)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.20)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.02)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.96)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.96)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.96)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.55)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.15)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.12)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.25)
33% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 455
\-------------------------

Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4033; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.53)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.83)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.26)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.37)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.18)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.02)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.82)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.82)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.71)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.50)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.10)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.12)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.27)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.45)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove right instead of left. (rewarded 0.93)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.73)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.62)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.62)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.67)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.88)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.88)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.88)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.53)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.93)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 456
\-------------------------

Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4025; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.20)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.36)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.36)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.36)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.36)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 0.99)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.21)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.12)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.41)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.41)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.41)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.41)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.02)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.02)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.26)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.26)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.26)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.26)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.24)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.91)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 457
\-------------------------

Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4017; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.02)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.52)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent followed the waypoint right. (rewarded 2.24)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 1.08)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.18)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.56)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.23)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.28)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.28)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.28)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.67)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.41)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.41)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.41)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.41)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.41)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.57)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.51)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.22)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.22)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.22)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.58)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 458
\-------------------------

Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4009; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.50)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.14)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.51)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.98)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.49)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.00)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.82)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.12)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.12)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.51)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.16)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.94)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.05)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.39)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.39)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.39)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.39)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.56)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.26)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.49)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 459
\-------------------------

Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.4001; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.84)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 1.27)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.32)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.67)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.67)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.60)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.63)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.14)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.14)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.14)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.58)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.50)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.50)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.10)
63% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.10)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded 1.20)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.28)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.09)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 0.76)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.99)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.74)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.08)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.93)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.93)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.93)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.93)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.93)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.88)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.37)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.37)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.37)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.37)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.95)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.23)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.41)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 460
\-------------------------

Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3993; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.39)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.61)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.79)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.74)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.29)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.46)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.46)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.46)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.46)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.46)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.46)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.83)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.33)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.69)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.54)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.90)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.90)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.46)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.29)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.29)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.27)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.85)
16% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 461
\-------------------------

Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3985; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.74)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.93)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.56)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.82)
86% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.82)
86% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.82)
86% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.82)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 1.12)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent drove left instead of right. (rewarded 0.08)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.16)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.33)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.33)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.33)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.33)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.33)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.72)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.00)
69% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.00)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.06)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.98)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.82)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.82)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.82)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.20)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.03)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.22)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.87)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.46)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.67)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.94)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.94)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.94)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded -0.03)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded -0.21)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.87)
29% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.87)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.08)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.08)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.13)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.12)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 0.44)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.15)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.03)
6% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.03)
6% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.03)
6% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.03)
6% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.03)
6% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.03)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.36)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.36)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 462
\-------------------------

Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3977; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.07)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.64)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -9.45)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -9.45)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -9.45)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -9.45)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -9.45)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.87)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.86)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.59)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.76)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.46)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.46)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.46)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.75)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.93)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.97)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.00)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.73)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.74)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.61)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.09)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.84)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.42)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.05)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 463
\-------------------------

Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3969; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.37)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.19)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.64)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.87)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.70)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.08)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.02)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.10)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.10)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.17)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.53)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 464
\-------------------------

Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3961; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.24)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.29)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.29)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.29)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.29)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.47)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.63)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 0.96)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.06)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.06)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.06)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.96)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 465
\-------------------------

Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3953; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.97)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.76)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.43)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.43)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.43)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.06)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.58)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.10)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.58)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.58)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.58)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.58)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.58)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.05)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.25)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.32)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.66)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.79)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.65)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.03)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.41)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 0.92)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.16)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.36)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.36)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.36)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.36)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.41)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.29)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.29)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.68)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.55)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.10)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.12)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 466
\-------------------------

Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3946; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.35)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.44)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.89)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.51)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.54)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.54)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.71)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.71)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.71)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.71)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.81)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.84)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.70)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.55)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.24)
49% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 467
\-------------------------

Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3938; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.59)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.73)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.26)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.26)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.82)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.40)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.78)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.96)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.92)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.92)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.14)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.67)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.93)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.24)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent followed the waypoint right. (rewarded 1.45)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 0.94)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 468
\-------------------------

Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3930; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.03)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent followed the waypoint left. (rewarded 2.59)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent followed the waypoint left. (rewarded 2.59)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.24)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.24)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.69)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.13)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.13)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.13)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.38)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.64)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.68)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.67)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.15)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 469
\-------------------------

Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3922; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.87)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.63)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.00)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.73)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.53)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.96)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.43)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.43)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.43)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.43)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.43)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 0.72)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.18)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.38)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.38)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.38)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.38)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.38)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.57)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.92)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.92)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.92)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.92)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.92)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.92)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.33)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.45)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.71)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.95)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.10)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.10)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.10)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.51)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.07)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.45)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.58)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.96)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.70)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.65)
17% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 470
\-------------------------

Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3914; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.08)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.70)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.89)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.89)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.82)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.07)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.86)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.76)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.92)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.24)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.04)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.66)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.25)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.73)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.72)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.40)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.40)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.83)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.60)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.60)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.60)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.86)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 471
\-------------------------

Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3906; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.33)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.80)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.12)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.87)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.08)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.38)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.75)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.24)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.24)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.24)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.35)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.94)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.01)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.01)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.01)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.28)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.93)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 472
\-------------------------

Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3898; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.60)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.02)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.76)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.12)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.27)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.58)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.14)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.18)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.40)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.13)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.71)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.71)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.71)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.75)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove right instead of forward. (rewarded 0.18)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.79)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.05)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded -0.36)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 473
\-------------------------

Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3891; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.05)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.66)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.66)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.66)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.11)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.09)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.99)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.29)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.88)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.88)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.88)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.89)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.89)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.12)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.69)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 2.63)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.43)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.88)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.23)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 0.83)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.34)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.67)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.67)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.67)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.25)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.86)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.87)
8% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 474
\-------------------------

Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3883; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.95)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.36)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.95)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 1.35)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 1.35)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 1.35)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 1.35)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 1.35)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.48)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.39)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.59)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.99)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.87)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.02)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.58)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.44)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.03)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 1.21)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.59)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.63)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.15)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.75)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 475
\-------------------------

Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3875; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.86)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.65)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.42)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.24)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.24)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.92)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.11)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.43)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.85)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.68)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.68)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.68)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.68)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.46)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.74)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.94)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.85)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 476
\-------------------------

Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3867; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.72)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.37)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.33)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.15)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.53)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.53)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.53)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.21)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.61)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.88)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.88)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.88)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.88)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.88)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.66)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.36)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.99)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.99)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.99)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.99)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.89)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.04)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.59)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.59)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.59)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.59)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.59)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.59)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.18)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.53)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.53)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.61)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.71)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.65)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.65)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.65)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.65)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.65)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.39)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 477
\-------------------------

Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3860; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.42)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.86)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.10)
91% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.10)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.28)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.29)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.48)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.48)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.48)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.48)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.02)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.68)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.72)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.54)
71% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.54)
71% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.54)
71% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.54)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.51)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.65)
66% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 478
\-------------------------

Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3852; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 0.12)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.68)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.08)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.72)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.56)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.25)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.12)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.41)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.41)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.50)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.50)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.15)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.87)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.70)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 479
\-------------------------

Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3844; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.24)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.57)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.17)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.53)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.27)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.19)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.19)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.19)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.81)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.09)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.09)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.09)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.13)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.66)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.66)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.66)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.47)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.01)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 480
\-------------------------

Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3837; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.68)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.31)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.07)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.95)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.15)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.47)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.63)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 481
\-------------------------

Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3829; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.75)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.40)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.41)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.52)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.52)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.11)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.56)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.27)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.65)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.89)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.66)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.34)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.34)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.34)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.34)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.34)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.34)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.35)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.17)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.84)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove left instead of right. (rewarded -0.13)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.94)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.33)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.33)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.33)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.84)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.76)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 482
\-------------------------

Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3821; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.37)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.43)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.67)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.68)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.71)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.47)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.35)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.29)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.53)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.92)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.10)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.10)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.12)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.61)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.61)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.61)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.07)
36% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 483
\-------------------------

Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3814; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.52)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.09)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.85)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.85)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.85)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.85)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.85)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.85)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.85)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.84)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.82)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.35)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 1.33)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.34)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 484
\-------------------------

Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3806; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.38)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.17)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.91)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.60)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.40)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.44)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.18)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.10)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.60)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 485
\-------------------------

Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3798; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.09)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.31)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.31)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.31)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.31)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.63)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.91)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.28)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.79)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 486
\-------------------------

Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3791; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.10)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.63)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.16)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.28)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.24)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.03)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.14)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.14)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.09)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.09)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.09)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.09)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.07)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.93)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.93)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.93)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.93)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.93)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.91)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.22)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.33)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.22)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.22)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.22)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.22)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.22)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.66)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.59)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.16)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.82)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.82)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.05)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 487
\-------------------------

Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3783; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.80)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.64)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.19)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.73)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.85)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.94)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.69)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.32)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.32)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.32)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.61)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 488
\-------------------------

Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3776; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.78)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.97)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.59)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.07)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.07)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.89)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.73)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.38)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.07)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.65)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.65)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.65)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.65)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.65)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.68)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.03)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.03)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.09)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.24)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.63)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.16)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.25)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.63)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.79)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.79)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.79)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'right')
Agent followed the waypoint right. (rewarded 0.77)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.58)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.40)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 489
\-------------------------

Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3768; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.46)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.75)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.00)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.29)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.29)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.06)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.17)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.83)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.23)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.23)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.23)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.23)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.11)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.83)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.79)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.79)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.79)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.79)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.81)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.51)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.23)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.27)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.27)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.27)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.26)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded -0.01)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.68)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.68)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.68)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.68)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.22)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.34)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.50)
8% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 490
\-------------------------

Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3761; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.41)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.54)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.31)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.31)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.31)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.31)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.35)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.35)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.35)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.29)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.33)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.52)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.63)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.63)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.63)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.63)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.71)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.91)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.44)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.97)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.69)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
43% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
43% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.03)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.21)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.21)
33% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.21)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded -0.45)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.14)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.14)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.06)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.44)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 1.22)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 1.22)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 1.22)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 1.22)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 1.22)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.94)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.27)
3% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 491
\-------------------------

Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3753; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.73)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.86)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.02)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.35)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.10)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.86)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.81)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.81)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.32)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.76)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.37)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.72)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.72)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.73)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.73)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.73)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.73)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.73)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.09)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.77)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.77)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.12)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.20)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 492
\-------------------------

Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3746; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.58)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.49)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.80)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.80)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.63)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.63)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.21)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.21)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.21)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.21)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.63)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.63)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.83)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.68)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.66)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 493
\-------------------------

Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3738; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.99)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.29)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.39)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.12)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.12)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.12)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.12)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.12)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.56)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.92)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.06)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.24)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.87)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.49)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.45)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.25)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 494
\-------------------------

Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3731; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.17)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.99)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.27)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.05)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.46)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.39)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.98)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.50)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.20)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.20)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.53)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.49)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.89)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.14)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.26)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.90)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.33)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.00)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 495
\-------------------------

Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3723; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 2.81)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.76)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.71)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.79)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.80)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.02)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.02)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.46)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.46)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.67)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.45)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.45)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.44)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 496
\-------------------------

Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3716; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.37)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'right', 'left')
Agent drove right instead of left. (rewarded 0.74)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 1.00)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 1.00)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 1.00)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 1.00)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 1.00)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.09)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.72)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.72)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.56)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.23)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.99)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.11)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.62)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.39)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.33)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.54)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 497
\-------------------------

Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3708; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.14)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.12)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.97)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.97)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.97)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.97)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.47)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded 1.49)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.77)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.38)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.02)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.64)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.29)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.47)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.22)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.06)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.88)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.39)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 0.60)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 498
\-------------------------

Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3701; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.94)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 1.70)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.96)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.96)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.96)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.96)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.96)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.79)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.27)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.49)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.64)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.64)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.64)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.32)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.18)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.76)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.96)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 499
\-------------------------

Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3694; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.56)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.31)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.69)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.31)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.23)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.76)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.23)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.23)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.46)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.68)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.04)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.04)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.86)
48% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.86)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.68)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.89)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.07)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.11)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.11)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.11)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.22)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.22)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.22)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'right', 'left')
Agent drove right instead of left. (rewarded -0.28)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded -0.26)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.23)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.94)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 500
\-------------------------

Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3686; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.34)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.84)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.40)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.88)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.24)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.22)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.93)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.16)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.16)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.56)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.82)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.99)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 501
\-------------------------

Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3679; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.67)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.16)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.16)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.90)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.39)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 502
\-------------------------

Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3671; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.42)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.36)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.62)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.95)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.24)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.24)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.24)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.08)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.11)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.16)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.16)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.16)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.16)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.16)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.45)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.81)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.27)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.10)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.33)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.22)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.91)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.89)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.43)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.43)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'right')
Agent drove right instead of forward. (rewarded 0.61)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.70)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 503
\-------------------------

Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3664; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.36)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.17)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.10)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.18)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.42)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.95)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.02)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.63)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.37)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 504
\-------------------------

Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3657; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.88)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.02)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.77)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.46)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.35)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 0.22)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.05)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.22)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.31)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.81)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.81)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.81)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.81)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.81)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.33)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.03)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.55)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.55)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.55)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.55)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.55)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.99)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.41)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 0.93)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.40)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.62)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 505
\-------------------------

Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3649; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.74)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.31)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.76)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.76)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.76)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.76)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.76)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.39)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.39)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.39)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.45)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent followed the waypoint right. (rewarded 2.21)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.62)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.62)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.62)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.12)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.12)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.53)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.89)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.16)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.16)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.35)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.54)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded -0.29)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.94)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 506
\-------------------------

Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3642; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent followed the waypoint forward. (rewarded 1.27)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.29)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.02)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.95)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.44)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.04)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent properly idled at a red light. (rewarded 0.50)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 0.98)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.89)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.33)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.19)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.61)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 507
\-------------------------

Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3635; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.28)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.80)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.82)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.90)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.90)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.90)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.42)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.64)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 0.51)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 1.21)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 508
\-------------------------

Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3628; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.44)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.57)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.90)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.42)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.42)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.42)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.42)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.01)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 1.81)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 509
\-------------------------

Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3620; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.47)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.48)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.07)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.45)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent drove right instead of left. (rewarded -0.01)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.61)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.34)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.29)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 510
\-------------------------

Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3613; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.80)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.98)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.90)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.31)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.29)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.45)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 511
\-------------------------

Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3606; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.78)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.69)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.22)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.83)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.13)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.68)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.23)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.74)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.28)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.90)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.28)
57% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 512
\-------------------------

Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3599; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.45)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.45)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.45)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.45)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.01)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.19)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.19)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.19)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.04)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.76)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent drove forward instead of left. (rewarded 1.50)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent drove forward instead of left. (rewarded 1.50)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent drove forward instead of left. (rewarded 1.50)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent drove forward instead of left. (rewarded 1.50)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.82)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.66)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.05)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.70)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.93)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.30)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.72)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.72)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.72)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.49)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 513
\-------------------------

Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3592; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.46)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.23)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.13)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.13)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.72)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.42)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.49)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.52)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.13)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.13)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.50)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.79)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.99)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.64)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.20)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.14)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.74)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.08)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.37)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.72)
43% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 514
\-------------------------

Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3584; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.05)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.75)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.96)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.50)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.14)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.18)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.40)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.18)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded -0.14)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.16)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.09)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.09)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.09)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.38)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.45)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.45)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.87)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 515
\-------------------------

Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3577; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.41)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.69)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.48)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.72)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded -0.01)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded -0.01)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded -0.01)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded -0.01)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded -0.01)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 0.22)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.39)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.39)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.39)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.54)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.69)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 0.51)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.58)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.58)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.58)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.99)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.99)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.99)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 516
\-------------------------

Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3570; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.92)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.75)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.28)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.62)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.62)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.62)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.62)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.84)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.51)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.11)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.18)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.18)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.18)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.18)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.18)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.42)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.59)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.01)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.04)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.04)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.04)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.28)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.17)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.41)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.41)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.41)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.41)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.99)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 1.09)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.61)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.80)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.35)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.96)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.65)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 517
\-------------------------

Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3563; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.49)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.53)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.88)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.90)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.25)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.79)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 518
\-------------------------

Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3556; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.04)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.22)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.58)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.81)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.81)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.81)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.20)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.07)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.46)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.46)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.45)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.69)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.76)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.19)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 0.76)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.62)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.62)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.62)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.62)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.42)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.63)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.14)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 519
\-------------------------

Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3549; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.33)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.46)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.38)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.38)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.38)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.38)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.42)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.42)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.42)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.42)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.15)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.25)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.78)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.78)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.78)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.65)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.32)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.45)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 520
\-------------------------

Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3542; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.06)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.33)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.35)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.91)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.45)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.10)
77% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 521
\-------------------------

Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3535; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.31)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.97)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.74)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.39)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.54)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.92)
72% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 522
\-------------------------

Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3527; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.78)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.12)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.16)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.32)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.61)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 2.77)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.24)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.61)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 1.39)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.15)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.21)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.56)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.11)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.95)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.93)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent drove right instead of left. (rewarded -0.50)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent drove right instead of left. (rewarded -0.50)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.05)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.30)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 523
\-------------------------

Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3520; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.22)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.60)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 1.29)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.64)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.55)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.52)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.82)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.58)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.04)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.10)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.35)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.44)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.26)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.42)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.42)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 0.86)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.47)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.34)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.22)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 0.99)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.45)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.45)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.45)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.23)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.57)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 524
\-------------------------

Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3513; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.52)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.34)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove left instead of forward. (rewarded 0.03)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove left instead of forward. (rewarded 0.03)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove left instead of forward. (rewarded 0.03)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove left instead of forward. (rewarded 0.03)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove left instead of forward. (rewarded 0.03)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.28)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.90)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.83)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.85)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 525
\-------------------------

Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3506; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.71)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.16)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.47)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.65)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.52)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.86)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.27)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.03)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 526
\-------------------------

Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3499; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.51)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.63)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.57)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.57)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.57)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.13)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.12)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.46)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.46)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.46)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.02)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.44)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.13)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.60)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 0.21)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.59)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 0.89)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.98)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.23)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 527
\-------------------------

Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3492; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.70)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.12)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.58)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.05)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.05)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.05)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.05)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.88)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.38)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.38)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.38)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.80)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.80)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.80)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.80)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.80)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.78)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.60)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.75)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.83)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.09)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.97)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.50)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.82)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 528
\-------------------------

Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3485; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.57)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.80)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.49)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.40)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.54)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.36)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.06)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.89)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.02)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.56)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.76)
48% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.76)
48% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.76)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.39)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.72)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.00)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded -0.45)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.56)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.86)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.13)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.13)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.13)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.13)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.49)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.27)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 0.68)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 529
\-------------------------

Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3478; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.57)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.88)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.12)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.34)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 530
\-------------------------

Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3471; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.43)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.44)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.06)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.52)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.52)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.52)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.52)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.43)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.14)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.18)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.35)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.26)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.91)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.60)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.51)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.51)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.51)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.51)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.51)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.56)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.24)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.59)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.29)
43% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 531
\-------------------------

Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3465; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.41)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.19)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.53)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.60)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.10)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.74)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.20)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.19)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.19)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.19)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.19)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.54)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.47)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.09)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.13)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.05)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.39)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.05)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.32)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.32)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.56)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.56)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.89)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.41)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 0.62)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 0.62)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.61)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.05)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.75)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 532
\-------------------------

Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3458; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.50)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.61)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.61)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.61)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.64)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.87)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.87)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.40)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.40)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.31)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.31)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.73)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.50)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.50)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.47)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.51)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.29)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.12)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.54)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 533
\-------------------------

Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3451; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.85)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.81)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.44)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.10)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.54)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.54)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.54)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.54)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.54)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.12)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.12)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.12)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.43)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.71)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.28)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.59)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.59)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.59)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.59)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.66)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.08)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.59)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.43)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.70)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.55)
5% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 534
\-------------------------

Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3444; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.50)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.01)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.48)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.79)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.79)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.82)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.82)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.49)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 0.03)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.30)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.30)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.41)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.41)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.41)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.41)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.41)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.03)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.31)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.31)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.46)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.51)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.04)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.88)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded -0.22)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'right')
Agent drove left instead of right. (rewarded -0.13)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 535
\-------------------------

Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3437; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.60)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.40)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.29)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.19)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.05)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.05)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.05)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.05)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.87)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.85)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.99)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.02)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded -0.18)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.13)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 0.98)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded -0.09)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.23)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.36)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.01)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.01)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.37)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 536
\-------------------------

Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3430; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.22)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.62)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.25)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.71)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.71)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.71)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.54)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.10)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.69)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.55)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.19)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.19)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.19)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.19)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.17)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.85)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.93)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.18)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.65)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.36)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.10)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.10)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.10)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.10)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.34)
24% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 537
\-------------------------

Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3423; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.24)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent drove right instead of forward. (rewarded 1.07)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.67)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.97)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.27)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.27)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.19)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.60)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.75)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 2.78)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 538
\-------------------------

Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3416; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.90)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.23)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.83)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.89)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.21)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.21)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.21)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.21)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.21)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.68)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.31)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.16)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.35)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.83)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.69)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 539
\-------------------------

Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3410; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 0.42)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 1.27)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.35)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.35)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.15)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.15)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.15)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.15)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.39)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.73)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.64)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.31)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.72)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.86)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.86)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.86)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.86)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.76)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.45)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.28)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.69)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 540
\-------------------------

Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3403; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.47)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.48)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 2.03)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.92)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.55)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.72)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.28)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.58)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 541
\-------------------------

Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3396; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.58)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.13)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.01)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.01)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.01)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.82)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.31)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.15)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.04)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.88)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.15)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded -0.06)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.26)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.26)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.26)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.41)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove left instead of right. (rewarded 0.42)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove left instead of right. (rewarded 0.42)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove left instead of right. (rewarded 0.42)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.72)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.02)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.93)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.93)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.91)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.59)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.98)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.98)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.98)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.98)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.51)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 542
\-------------------------

Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3389; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.06)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.72)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.14)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.35)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.61)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.48)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.12)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.62)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.62)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.62)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.91)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.91)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.91)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.64)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.19)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.79)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded -0.19)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded -0.19)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.33)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 543
\-------------------------

Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3382; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.66)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.69)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.84)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.25)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.52)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.95)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 1.83)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 544
\-------------------------

Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3376; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.78)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.42)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.94)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.89)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.62)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.35)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.36)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.42)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.28)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.12)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.12)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.12)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.12)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.12)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.50)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.76)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.27)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.71)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 545
\-------------------------

Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3369; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.49)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.38)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.33)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.33)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.33)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.33)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.33)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.33)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.45)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.45)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.64)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 546
\-------------------------

Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3362; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent drove forward instead of right. (rewarded 1.02)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.40)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.45)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.54)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.54)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.54)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.54)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.44)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.28)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.04)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.97)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 547
\-------------------------

Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3355; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.16)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.54)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.30)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.92)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.92)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.92)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.61)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.87)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.13)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.01)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.01)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.01)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.41)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.76)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.50)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.50)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.50)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.90)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.90)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.90)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.16)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.54)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 2.15)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 548
\-------------------------

Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3349; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.23)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.35)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.92)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.96)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.38)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.16)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.38)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.57)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.90)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.05)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.05)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.05)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.43)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.73)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.14)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.43)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 549
\-------------------------

Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3342; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.65)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.29)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.61)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.76)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.43)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.16)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.34)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.78)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.11)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.74)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.82)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 550
\-------------------------

Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3335; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.60)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.48)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.93)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.18)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.12)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.33)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent drove forward instead of left. (rewarded 0.42)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.84)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.84)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.84)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.84)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.73)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 551
\-------------------------

Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3329; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.35)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.95)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.45)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.39)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.39)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.32)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.23)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.33)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.38)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.37)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.40)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.04)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.04)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.78)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.43)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.43)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.43)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.26)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.58)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.50)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.08)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.08)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded -0.31)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 0.34)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.66)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 552
\-------------------------

Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3322; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.44)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.93)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.20)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.20)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.28)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.59)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.07)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.67)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 553
\-------------------------

Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3315; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.64)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.33)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.57)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.30)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.24)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.84)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.73)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.91)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.54)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.54)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.54)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.54)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.75)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.80)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.25)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.25)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.25)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.25)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.25)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.51)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.26)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.26)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.75)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.75)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove forward instead of left. (rewarded -0.54)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.42)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.71)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 554
\-------------------------

Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3309; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.14)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.80)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.63)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.63)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.63)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 0.96)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.37)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 555
\-------------------------

Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3302; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.66)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.37)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.59)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.70)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.45)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.90)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 556
\-------------------------

Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3296; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.53)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.49)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.64)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.05)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.94)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.21)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.21)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.21)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.91)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.94)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.94)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.94)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.83)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.58)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.06)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 557
\-------------------------

Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3289; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.37)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.64)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.41)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.18)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.61)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.10)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.70)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent drove forward instead of left. (rewarded 0.68)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.10)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.70)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.70)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.70)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.70)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.56)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.79)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.79)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.29)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.69)
47% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 558
\-------------------------

Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3282; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.65)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.72)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.42)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.67)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.13)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 559
\-------------------------

Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3276; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.96)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.55)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.02)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.21)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.06)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.06)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.57)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.69)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 1.27)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.94)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.23)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.51)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.82)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.84)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 560
\-------------------------

Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3269; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.13)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.01)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.95)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.95)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.95)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.95)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.95)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.06)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.14)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.17)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.05)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.01)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.01)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.01)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.01)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.84)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.84)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.80)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.49)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.28)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.53)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.53)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.53)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.63)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.22)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.07)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.18)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.18)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.42)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.42)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.42)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.50)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.72)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.92)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 561
\-------------------------

Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3263; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.94)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.25)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.88)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.29)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.18)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.70)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.69)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.69)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.69)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.69)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.04)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 0.52)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 0.52)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.52)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.52)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.81)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.84)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.22)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.22)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.22)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.16)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.15)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.15)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.15)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.34)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.58)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.58)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.99)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 562
\-------------------------

Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3256; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.29)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.31)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.02)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.28)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 0.44)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.76)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.19)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.25)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.93)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.12)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.12)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.12)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.78)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.50)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.60)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.79)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.16)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.75)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.75)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.53)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 563
\-------------------------

Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3250; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.97)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.96)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.22)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.22)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.63)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.41)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.84)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 564
\-------------------------

Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3243; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.94)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.26)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.62)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.78)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.93)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.50)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.28)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.03)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.99)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.23)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.48)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.55)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.55)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.55)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.55)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.55)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.55)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.55)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.55)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.47)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 565
\-------------------------

Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3237; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.88)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.19)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.14)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.14)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.14)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.14)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.51)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.54)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 2.71)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 2.71)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.06)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.20)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.20)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.20)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.20)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.20)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.20)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.38)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.29)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.70)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.88)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.10)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.10)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.10)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 1.25)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.40)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.14)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.14)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.14)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.14)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.42)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.42)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.42)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.99)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 0.81)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.16)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.66)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.66)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.46)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 566
\-------------------------

Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3230; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.40)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.04)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.53)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.53)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.53)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.53)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.91)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.80)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.02)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.98)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.76)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.56)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.03)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.25)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 567
\-------------------------

Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3224; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.04)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.12)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.58)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.13)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.97)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.72)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.64)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.72)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.51)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.54)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.20)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.38)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.38)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.55)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.45)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 568
\-------------------------

Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3217; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.45)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.85)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.85)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.85)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.85)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.21)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.60)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.46)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.46)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.46)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.46)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.26)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.65)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.65)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.65)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.65)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.62)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.14)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.14)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.14)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.08)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.50)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.71)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.28)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.95)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.88)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.32)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.35)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.89)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.45)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.45)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.45)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.45)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.45)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.64)
13% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 569
\-------------------------

Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3211; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'left')
Agent followed the waypoint left. (rewarded 2.12)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.45)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.51)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.74)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.74)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.54)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.86)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.58)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 570
\-------------------------

Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3205; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.02)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.87)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.67)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.98)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.98)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.98)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.98)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.98)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.57)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.70)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.64)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.25)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.94)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 0.44)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 0.44)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.54)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.13)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.19)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.42)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.42)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.48)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.20)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.14)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.51)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 571
\-------------------------

Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3198; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.91)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.88)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.94)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.67)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.32)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.05)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.54)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.43)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 572
\-------------------------

Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3192; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.91)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.37)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.45)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.45)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.05)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.05)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.05)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.76)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.91)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.11)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.11)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.51)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.11)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.37)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.61)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.97)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.97)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.97)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.97)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.00)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.72)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.42)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.81)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.57)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 573
\-------------------------

Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3185; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.16)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.23)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 1.12)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.55)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.55)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.89)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'right')
Agent followed the waypoint forward. (rewarded 1.58)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.49)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.49)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.49)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.49)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.49)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.43)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.59)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.59)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.59)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.60)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 574
\-------------------------

Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3179; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.51)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.53)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.76)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.23)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.93)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 1.93)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 575
\-------------------------

Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3173; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.08)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.31)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.31)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.21)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.21)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.88)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.88)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.88)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.88)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.88)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.88)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.88)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.28)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.20)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.44)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.98)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.27)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.34)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.20)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 576
\-------------------------

Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3166; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.30)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.96)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.22)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.73)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.99)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.83)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.01)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.02)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 577
\-------------------------

Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3160; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.81)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.62)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.41)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.41)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.41)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.41)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.41)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.41)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.41)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.65)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.66)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.16)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.82)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 1.69)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 1.69)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 1.69)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 1.69)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.03)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.77)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 578
\-------------------------

Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3154; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.13)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.31)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.63)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.63)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.63)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.63)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.59)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.19)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.19)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.19)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.19)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.19)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.96)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.23)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.53)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.91)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 579
\-------------------------

Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3147; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 1.74)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.17)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.69)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.56)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 580
\-------------------------

Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3141; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.05)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.19)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.25)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.69)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.21)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.41)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.85)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.06)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.91)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.91)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.92)
57% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 581
\-------------------------

Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3135; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 0.42)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.51)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.00)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.00)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.00)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.36)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.07)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.87)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.64)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.72)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.94)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.94)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.94)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.94)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.40)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.40)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.40)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.40)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.57)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.84)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.84)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.90)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.90)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.90)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.92)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.12)
43% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 582
\-------------------------

Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3129; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.57)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.85)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.04)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.31)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.18)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.33)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.48)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.48)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.48)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.35)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 583
\-------------------------

Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3122; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.26)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.33)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.98)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.72)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.16)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.98)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.60)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.63)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.23)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.90)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.30)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.79)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.90)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.76)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.76)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.76)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.76)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.45)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.45)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.11)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.60)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.14)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.05)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.05)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.05)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.27)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.27)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.27)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.39)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.94)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.58)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.29)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.31)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 584
\-------------------------

Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3116; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.91)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.94)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.32)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.18)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.65)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.74)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.43)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.96)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 585
\-------------------------

Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3110; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove forward instead of left. (rewarded 0.66)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.38)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.81)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.81)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.81)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.81)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.81)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.81)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.83)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.83)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.63)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.70)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.73)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.66)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 586
\-------------------------

Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3104; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.85)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.72)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.27)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.43)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.15)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.65)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 587
\-------------------------

Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3097; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', None)
Agent drove right instead of left. (rewarded 0.62)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.23)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.69)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 1.03)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 588
\-------------------------

Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3091; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.57)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.22)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.70)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.30)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.64)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.39)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.75)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.28)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.31)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.81)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.60)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 589
\-------------------------

Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3085; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.63)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.67)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.93)
89% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.93)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.22)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.51)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.17)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.17)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.36)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.36)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.65)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.70)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.96)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent drove left instead of right. (rewarded 1.68)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.24)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.23)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.62)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.72)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.80)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.00)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.12)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.97)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.39)
34% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.39)
34% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.39)
34% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.39)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.29)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.46)
29% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.46)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.33)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.02)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.50)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.50)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.50)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.35)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 0.28)
14% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 0.28)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.68)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.94)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.29)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.26)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 590
\-------------------------

Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3079; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.75)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.73)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.64)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.03)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.03)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.03)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.50)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'right')
Agent drove left instead of forward. (rewarded 0.22)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.78)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 0.94)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.67)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.88)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 591
\-------------------------

Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3073; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.29)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.85)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 1.58)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 1.58)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 1.58)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.81)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.04)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.04)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.04)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.04)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.01)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.60)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.92)
67% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 592
\-------------------------

Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3067; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.70)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.89)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.61)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.61)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.30)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.64)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.64)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.64)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.64)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.70)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.91)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.95)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.25)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.25)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.25)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.25)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.79)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.12)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.12)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.12)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.12)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.12)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.30)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.57)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.25)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.59)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.33)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 593
\-------------------------

Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3061; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent followed the waypoint right. (rewarded 2.60)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.73)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.20)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.80)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.80)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.87)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.65)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.20)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 594
\-------------------------

Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3054; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.12)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.92)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.77)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 595
\-------------------------

Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3048; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.11)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.31)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.72)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.81)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.81)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.68)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.68)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.68)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.41)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.41)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.41)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.92)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.92)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.16)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.50)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 596
\-------------------------

Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3042; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.47)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.06)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.59)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.76)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.92)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.41)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.04)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.57)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.57)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.57)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.57)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 1.60)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.56)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.41)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.34)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.04)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.20)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.20)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.20)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.20)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.20)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.02)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 597
\-------------------------

Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3036; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.88)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.53)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.25)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.67)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.48)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.48)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.48)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.48)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.68)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.49)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.49)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.43)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.43)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.43)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.43)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 1.66)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 598
\-------------------------

Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3030; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.11)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.75)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.90)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.90)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.48)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.48)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.52)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.52)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.52)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.52)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.59)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.71)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 599
\-------------------------

Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3024; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.06)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.65)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.47)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'right')
Agent drove right instead of forward. (rewarded 0.27)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.84)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.92)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.89)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.15)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.39)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.26)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.09)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.46)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 0.94)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.34)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.27)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.34)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.83)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.14)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.85)
4% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 600
\-------------------------

Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3018; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.02)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.39)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.71)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.38)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.60)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.07)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.40)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.40)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.40)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.40)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.40)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.08)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.29)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.29)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.29)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.12)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.27)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded -0.62)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.55)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.15)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 601
\-------------------------

Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3012; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 1.41)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.92)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.42)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.79)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.94)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.90)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.20)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.16)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.65)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.98)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.98)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.98)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.98)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.82)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.35)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent followed the waypoint right. (rewarded 1.69)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.83)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.58)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.09)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove left instead of right. (rewarded -0.32)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 602
\-------------------------

Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3006; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.95)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.97)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.03)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.03)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.92)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.92)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.16)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.38)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.80)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.13)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.02)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.70)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.08)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 603
\-------------------------

Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.3000; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.28)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.01)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.39)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.33)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.30)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 604
\-------------------------

Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2994; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.01)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.52)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.56)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 1.53)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.27)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.27)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.27)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.27)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.08)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.84)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.36)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.36)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded 1.40)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.98)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.39)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.39)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.86)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.86)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.86)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.86)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.22)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.90)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 605
\-------------------------

Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2988; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.12)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 2.11)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.52)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.05)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.07)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.44)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.36)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 606
\-------------------------

Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2982; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.72)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.84)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.66)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.42)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.94)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.18)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.94)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.08)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.35)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.30)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.28)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.44)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.55)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.48)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.48)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.53)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.99)
47% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.99)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.29)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded -0.20)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.80)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.41)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 607
\-------------------------

Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2976; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.93)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.63)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.19)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.48)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.28)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.15)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.15)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.15)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.15)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.15)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.36)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.20)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.29)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.59)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.35)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.77)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.00)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.68)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.68)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.68)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.53)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.89)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent drove right instead of forward. (rewarded 0.05)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.52)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.52)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.52)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.72)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.04)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.35)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.16)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.82)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.82)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.17)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded -0.53)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 608
\-------------------------

Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2970; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.04)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.86)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.54)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.54)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.54)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.54)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.67)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.26)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.94)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.83)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.12)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.24)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.24)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.24)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.44)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 609
\-------------------------

Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2964; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.57)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.53)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.88)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.88)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.88)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.42)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.89)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.76)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.06)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.67)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.88)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 610
\-------------------------

Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2958; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove forward instead of left. (rewarded 1.09)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.39)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.94)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.18)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.18)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.18)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.18)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.18)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.18)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.58)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.85)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.18)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.18)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.64)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.04)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.61)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.46)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.98)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.62)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.24)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.24)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.07)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.26)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.52)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.52)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.25)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.76)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.91)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.30)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.57)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.86)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.99)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 611
\-------------------------

Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2952; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.51)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.58)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.29)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.29)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.29)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.29)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.29)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.85)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.17)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.61)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.85)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.63)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 612
\-------------------------

Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2946; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.71)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.97)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.22)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.22)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.22)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.56)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded 1.20)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.78)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.74)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.71)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.95)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.57)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.44)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.44)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.44)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.44)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.61)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.72)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 613
\-------------------------

Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2941; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.88)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.25)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.78)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.37)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.92)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.94)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.66)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.66)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.66)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.66)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.85)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.95)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.95)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.95)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.60)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.22)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.22)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.22)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.22)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.10)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.10)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.40)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.49)
28% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.49)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.93)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.32)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.57)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.57)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.57)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.57)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.57)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.57)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.29)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.18)
8% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 614
\-------------------------

Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2935; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.63)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.42)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.33)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.35)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.35)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.91)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.12)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.27)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 615
\-------------------------

Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2929; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.53)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.74)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.93)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.28)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.64)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.64)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.64)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.90)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.73)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.32)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.13)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.46)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.46)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.46)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.54)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.00)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.00)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.00)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.00)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.00)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.00)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.66)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded -0.02)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.09)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.09)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.61)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.99)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 616
\-------------------------

Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2923; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.13)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.84)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.27)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.68)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.68)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.62)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.45)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.33)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.33)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.33)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.33)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent drove forward instead of left. (rewarded 0.31)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.74)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.82)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.35)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.35)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.11)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.36)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.34)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 617
\-------------------------

Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2917; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.68)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.69)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.69)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.69)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.69)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.88)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.94)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.98)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.60)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.60)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.60)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.60)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.60)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.60)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.79)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.39)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.39)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.40)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.40)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.46)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.46)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.17)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.10)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.10)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.10)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.10)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.28)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.72)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.68)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 618
\-------------------------

Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2911; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.33)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.84)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.49)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.79)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.79)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.30)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.17)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.17)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.17)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.33)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.33)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.40)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.40)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.93)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.79)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.94)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.75)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.97)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.97)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.97)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.97)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.97)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.97)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.96)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.18)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.43)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.74)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.91)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.27)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.27)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.13)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.84)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.84)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.84)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.84)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.84)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.78)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.97)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.97)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.97)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.97)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.81)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded -0.69)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.12)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 619
\-------------------------

Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2905; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.56)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.86)
93% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.86)
93% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.86)
93% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.86)
93% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.86)
93% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.86)
93% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.86)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.09)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.93)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.27)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.44)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.85)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.85)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.85)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.41)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.41)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.41)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.43)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.31)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.32)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded -0.11)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.08)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.08)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.08)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.13)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.22)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent drove right instead of left. (rewarded 1.61)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.27)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.27)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.40)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.58)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.58)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.58)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.58)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.23)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.74)
13% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.74)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.48)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.48)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.48)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.81)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.40)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 620
\-------------------------

Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2900; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.76)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.30)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.66)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.03)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.19)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.19)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.19)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.19)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.19)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.05)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.69)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.27)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.63)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.22)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.38)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.38)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.38)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.38)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.46)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.99)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.62)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.72)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.72)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.72)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.72)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.28)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.14)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.93)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.93)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.52)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 621
\-------------------------

Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2894; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.85)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.13)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.62)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.62)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.62)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.62)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.62)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.96)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.95)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.95)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.95)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.95)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.95)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.18)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.93)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.62)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.62)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.57)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.57)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.80)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.60)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.34)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.34)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.66)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.17)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.13)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent drove forward instead of left. (rewarded -0.20)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.37)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.87)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.17)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.56)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.21)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 622
\-------------------------

Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2888; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.11)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.76)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.90)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.23)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.34)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.88)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.88)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.88)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.45)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.26)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.51)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.67)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.75)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.75)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.75)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.69)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.69)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.27)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.69)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.46)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.18)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded -0.31)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded -0.31)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded -0.31)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.97)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.64)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.18)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.50)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.40)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.08)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.09)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 623
\-------------------------

Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2882; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.87)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.34)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.47)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.32)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.71)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.71)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.71)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.71)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.71)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.41)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.48)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.68)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.27)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.14)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.09)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.39)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.64)
24% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 624
\-------------------------

Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2877; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.38)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.21)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 1.27)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.03)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.03)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.03)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.03)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.03)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.30)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.07)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.69)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.49)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 625
\-------------------------

Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2871; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.34)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.72)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.38)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.38)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.38)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.38)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.38)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.11)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.96)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.24)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.39)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.67)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.88)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.85)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.36)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.80)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.21)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.29)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.65)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 626
\-------------------------

Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2865; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.67)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.53)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.95)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.83)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.30)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.77)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove right instead of left. (rewarded 0.67)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.17)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.26)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.29)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 627
\-------------------------

Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2859; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 2.00)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.78)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.68)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.68)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.56)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.20)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.20)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.20)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.20)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.20)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.96)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.86)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.98)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 0.78)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 0.78)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'right')
Agent followed the waypoint right. (rewarded 1.16)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.14)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.94)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.94)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.33)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.33)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.33)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.21)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.62)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.20)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.20)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.20)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.20)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.20)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded -0.43)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.43)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.43)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.43)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.43)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.43)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.65)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.74)
4% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 628
\-------------------------

Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2854; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'right')
Agent followed the waypoint right. (rewarded 1.47)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.43)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.17)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.68)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 0.09)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.87)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.17)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.17)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.17)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.55)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.09)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent drove right instead of left. (rewarded 1.54)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.04)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.04)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.64)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.39)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.98)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.24)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.18)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 629
\-------------------------

Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2848; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.35)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.29)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.72)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.72)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.72)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.16)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.05)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.05)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.87)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.87)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.69)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.57)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.62)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.51)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.45)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.45)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.45)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.45)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.45)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.45)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.81)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.55)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.55)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.55)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.55)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.55)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.55)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.31)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.57)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.04)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.18)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.70)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.81)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.56)
4% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 630
\-------------------------

Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2842; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.83)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.57)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.04)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.10)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.13)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.13)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.13)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.54)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.54)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.54)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.80)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.53)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.07)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 631
\-------------------------

Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2837; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.41)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.03)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.30)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.46)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 0.92)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 0.96)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 1.78)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 632
\-------------------------

Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2831; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.02)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.55)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.22)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.22)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.22)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.22)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.34)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.25)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.25)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.25)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.25)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.26)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.50)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.85)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.64)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.46)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.55)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.55)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.55)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.55)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.78)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.92)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.78)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded -0.39)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded -0.39)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded -0.39)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded -0.39)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.90)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded -0.48)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded -0.48)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded -0.48)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.53)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 633
\-------------------------

Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2825; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.60)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove left instead of forward. (rewarded 1.54)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.77)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.14)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.03)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.03)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.88)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.49)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.95)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.22)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.38)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 634
\-------------------------

Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2820; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.16)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.43)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.93)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.04)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.54)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.59)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.62)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.42)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.01)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 635
\-------------------------

Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2814; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.51)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.06)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.89)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.46)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.46)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.01)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.89)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.34)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 636
\-------------------------

Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2808; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 2.45)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.42)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.98)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.98)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.98)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.98)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.24)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.88)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.40)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.13)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 637
\-------------------------

Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2803; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.86)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.11)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.97)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.69)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.87)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.87)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.87)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.87)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.87)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.46)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 0.61)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.56)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.76)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.28)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.27)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.61)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded -0.30)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.42)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.83)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 638
\-------------------------

Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2797; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.35)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.78)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.78)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.78)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.78)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.98)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.23)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.42)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.86)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.16)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.98)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.62)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.62)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.43)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.89)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.36)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.36)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.36)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.36)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.36)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.08)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.04)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.04)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.65)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.25)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.49)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.49)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.49)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.49)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.49)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.04)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.00)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.91)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.14)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.09)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.86)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.78)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 639
\-------------------------

Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2792; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.97)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.24)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.02)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 640
\-------------------------

Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2786; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.97)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.17)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.29)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.29)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 0.85)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.90)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.38)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.38)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.38)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.38)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.38)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.56)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.63)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.07)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.11)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.11)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.11)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.11)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.36)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.54)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.76)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.76)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.93)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent drove left instead of forward. (rewarded -0.36)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.06)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 1.10)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded -0.35)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.46)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 641
\-------------------------

Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2780; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.54)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.45)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.78)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.69)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 1.09)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.42)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.82)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.33)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 642
\-------------------------

Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2775; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.39)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.92)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.49)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.22)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.31)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.19)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.81)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.32)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.26)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.26)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.26)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.26)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.14)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.84)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.84)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.84)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.84)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.30)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.96)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.27)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.24)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.15)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.03)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.55)
33% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 643
\-------------------------

Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2769; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.00)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.01)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.01)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.01)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.01)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.12)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.56)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.26)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.28)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 1.02)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.27)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', None)
Agent drove right instead of left. (rewarded 0.50)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.62)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.29)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.58)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.90)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.70)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.00)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.64)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 644
\-------------------------

Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2764; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.01)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.20)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.85)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.81)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.94)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.89)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.58)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 645
\-------------------------

Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2758; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.28)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.46)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.79)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.79)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.79)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.79)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.79)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.79)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.80)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.22)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.41)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.13)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.29)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.12)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.24)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 646
\-------------------------

Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2753; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.03)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.01)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.95)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.95)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.95)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.23)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.23)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.68)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.10)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.10)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.10)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.36)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.87)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.39)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.32)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.99)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.99)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.99)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.99)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.99)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.78)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.54)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.54)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.93)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.37)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.19)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.33)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.94)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.94)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.94)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.78)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.52)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 647
\-------------------------

Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2747; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.11)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.96)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.47)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.47)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.47)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.59)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.32)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.32)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.99)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 1.72)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 648
\-------------------------

Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2742; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.82)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.06)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.16)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.72)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.46)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 649
\-------------------------

Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2736; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.22)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.54)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.43)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.95)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.88)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.11)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.22)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.65)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.75)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent drove right instead of forward. (rewarded -0.28)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent drove right instead of forward. (rewarded -0.28)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent drove right instead of forward. (rewarded -0.28)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.22)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.44)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 0.88)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.93)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.09)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 650
\-------------------------

Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2731; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.04)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.72)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.81)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 0.95)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.11)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.41)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.68)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.94)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.11)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 651
\-------------------------

Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2725; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.21)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.79)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.16)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.09)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.09)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.55)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.31)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.31)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.31)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.31)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.91)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.73)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.73)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.73)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.73)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'right')
Agent drove forward instead of left. (rewarded 0.10)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 0.88)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 0.88)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.67)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.67)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.24)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.54)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.52)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded -0.07)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.39)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.50)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.50)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.50)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.71)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.19)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 652
\-------------------------

Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2720; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.42)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.23)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.80)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.80)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.80)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.80)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.80)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent drove right instead of left. (rewarded 0.83)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.14)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.25)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.35)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.84)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.64)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded -0.00)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.42)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.32)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.32)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.32)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.32)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.32)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.32)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.13)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.85)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.24)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.43)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.51)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 653
\-------------------------

Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2714; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.51)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.48)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.60)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.60)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.60)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.60)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.40)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.46)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.41)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.73)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.54)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.56)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.30)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.87)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 654
\-------------------------

Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2709; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.56)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.02)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.67)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.96)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.89)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.73)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.94)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.23)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 1.45)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.28)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove forward instead of left. (rewarded 1.45)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove forward instead of left. (rewarded 1.45)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove forward instead of left. (rewarded 1.45)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.42)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.75)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.45)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.74)
8% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 655
\-------------------------

Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2704; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.44)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.61)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.83)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.30)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.30)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.30)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.30)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.09)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.23)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.46)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.37)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.92)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.92)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.92)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.92)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.06)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.02)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.15)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.13)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.53)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.07)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.68)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.43)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.43)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.43)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.43)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.84)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.43)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.89)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.89)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.89)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.79)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.46)
26% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.46)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.00)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.37)
14% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 656
\-------------------------

Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2698; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.39)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.84)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.20)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.20)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.39)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.39)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.39)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.39)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.94)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.52)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.75)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 657
\-------------------------

Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2693; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.80)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.14)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.30)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.46)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.31)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.46)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.89)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.67)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.98)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.98)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.98)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.98)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.98)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.52)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.77)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.83)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.90)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.40)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.02)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.42)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.42)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.42)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.30)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.43)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.43)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.43)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.43)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.43)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.94)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.34)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.14)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 658
\-------------------------

Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2687; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.52)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.57)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.44)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.44)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.44)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.44)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.75)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.01)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.00)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.80)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.36)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.06)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.27)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.27)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.27)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.27)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.52)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.10)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 659
\-------------------------

Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2682; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.99)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.27)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.90)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.19)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.31)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.86)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded -0.03)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.23)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.28)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.34)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded -0.08)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent followed the waypoint right. (rewarded 1.61)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.72)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 660
\-------------------------

Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2677; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.86)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.49)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 2.46)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.41)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.23)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.23)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.71)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.71)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.71)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.71)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.59)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.58)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.51)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.97)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.88)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.41)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.38)
36% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 661
\-------------------------

Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2671; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.36)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.84)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.56)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.20)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.88)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 662
\-------------------------

Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2666; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.83)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.82)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.60)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.14)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 663
\-------------------------

Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2661; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.88)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.72)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.30)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.31)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.92)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.54)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.39)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 1.49)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.45)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.93)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 1.39)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.52)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.42)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.39)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 664
\-------------------------

Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2655; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.09)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.23)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.27)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.50)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.94)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.96)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.96)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.96)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.96)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.70)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.35)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.35)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.48)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.22)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.31)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.79)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 665
\-------------------------

Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2650; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.38)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.06)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.12)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.05)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.05)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.06)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.82)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.92)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.69)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.23)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.56)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.06)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.04)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.57)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 666
\-------------------------

Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2645; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.02)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.50)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.56)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.57)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.77)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.21)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.86)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.76)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.37)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.82)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.74)
69% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.74)
69% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.74)
69% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.74)
69% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.74)
69% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.74)
69% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.74)
69% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.74)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.09)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.09)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.72)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.72)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.72)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.72)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.72)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.69)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.86)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.45)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.07)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.81)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.53)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 0.90)
37% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 667
\-------------------------

Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2639; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.70)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.02)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.02)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.88)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.55)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.57)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.26)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 1.45)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.72)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.50)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.83)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.71)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.23)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.23)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.23)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.61)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.63)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.63)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.63)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.63)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.63)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.63)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.63)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 0.34)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 668
\-------------------------

Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2634; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.06)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.84)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.30)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 1.44)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.73)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.59)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.33)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.49)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.26)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.78)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.78)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.78)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.78)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.85)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.05)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.99)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.30)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.30)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 0.49)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 0.97)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.46)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 669
\-------------------------

Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2629; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.14)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.08)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 2.97)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.44)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.01)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 670
\-------------------------

Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2624; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.58)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.07)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.88)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.84)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.84)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.84)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.84)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.92)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.92)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.60)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.60)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.60)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.60)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.60)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.60)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 0.73)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.32)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.22)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove left instead of right. (rewarded 0.74)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove left instead of right. (rewarded 0.74)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove left instead of right. (rewarded 0.74)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.98)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.98)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.98)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.98)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.61)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded 0.16)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.73)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.29)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.95)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.47)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.19)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.28)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.39)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.87)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.39)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.90)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.17)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.87)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.87)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.87)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.87)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.87)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.93)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.59)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.83)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.36)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 671
\-------------------------

Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2618; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.21)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.34)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.46)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.80)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.80)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.80)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.80)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.80)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.80)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.96)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.93)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.16)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.11)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.11)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.11)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.14)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.43)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.47)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.62)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 0.82)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.80)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.82)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.82)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.82)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.82)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.82)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.82)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.82)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.61)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent drove forward instead of right. (rewarded 0.24)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent drove forward instead of right. (rewarded 1.18)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.17)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.17)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.17)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.17)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.17)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.17)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.17)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.76)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.07)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.33)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.29)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.02)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.58)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.12)
7% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 672
\-------------------------

Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2613; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.06)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.10)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.89)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.40)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.40)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.40)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.33)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.50)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.41)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.76)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.82)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.05)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.33)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 0.72)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.11)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.01)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.12)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.79)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.71)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.18)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.15)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.15)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.15)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.84)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 673
\-------------------------

Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2608; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.76)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.43)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.02)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.06)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.25)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.81)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 674
\-------------------------

Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2603; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.93)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.50)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.32)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.23)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.81)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.44)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.80)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 675
\-------------------------

Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2598; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 1.56)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.80)
92% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.80)
92% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.80)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.01)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.01)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.01)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.01)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.31)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 676
\-------------------------

Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2592; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.94)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.93)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.08)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.72)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.72)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.17)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.99)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.13)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.12)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.12)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.12)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.12)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.12)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.12)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.12)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.98)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.58)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.03)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.68)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.68)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.68)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.68)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.68)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.68)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.59)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.77)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.18)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.18)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.18)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.03)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 677
\-------------------------

Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2587; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.34)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.18)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.44)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.26)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 678
\-------------------------

Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2582; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.98)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.76)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.16)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.63)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.34)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.81)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.81)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.81)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.81)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.72)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.76)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.76)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.11)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.11)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.95)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.01)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded -0.13)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.37)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.57)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.58)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.76)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 1.10)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.38)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.74)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.40)
17% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 679
\-------------------------

Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2577; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.06)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.33)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.93)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.87)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.96)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.16)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.39)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.97)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.34)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.22)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 680
\-------------------------

Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2572; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.61)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.59)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.64)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.51)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.30)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.26)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.86)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.97)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.62)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove right instead of left. (rewarded -0.11)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.47)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.12)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.59)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.41)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.82)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.82)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.40)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.40)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.40)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.89)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 0.70)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.00)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.00)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.00)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.82)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'right')
Agent attempted driving left through a red light. (rewarded -9.41)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.37)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 681
\-------------------------

Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2567; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.44)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.63)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.10)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.10)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.10)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.10)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.10)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.36)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.72)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.52)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.57)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.57)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.57)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.57)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.42)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.97)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.65)
69% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.65)
69% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.65)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.94)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded -0.11)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded -0.11)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded -0.11)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded -0.11)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded -0.11)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.76)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.18)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.60)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.49)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.27)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.13)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.99)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove left instead of right. (rewarded 0.20)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.28)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.63)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.32)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 0.75)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 0.45)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 1.69)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.50)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
14% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
14% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.90)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.98)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.67)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.67)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.67)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.67)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.67)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.27)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 682
\-------------------------

Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2561; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.71)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.67)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.97)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.46)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.51)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.82)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.85)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 683
\-------------------------

Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2556; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.97)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.68)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.07)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.80)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.29)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.29)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.07)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.18)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.73)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.81)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.72)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 684
\-------------------------

Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2551; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.94)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.82)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.59)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.52)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.66)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.12)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.85)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.26)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.26)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.98)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.71)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.21)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.21)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.55)
57% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 685
\-------------------------

Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2546; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.12)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.82)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 1.26)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.72)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.62)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 686
\-------------------------

Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2541; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.80)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.90)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.84)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.82)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.82)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.82)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.82)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.40)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.35)
72% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 687
\-------------------------

Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2536; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.83)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.15)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.22)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.57)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.39)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.56)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.13)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.13)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 1.37)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 688
\-------------------------

Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2531; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.84)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.45)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.26)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.34)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.49)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.14)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.14)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.14)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.14)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.82)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.98)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.66)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.72)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.72)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.06)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.06)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.20)
36% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 689
\-------------------------

Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2526; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.15)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.36)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.82)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.90)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.04)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.33)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.24)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.52)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded -0.02)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 0.85)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded 1.61)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.80)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.82)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.05)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.05)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.89)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.89)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.68)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.82)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.19)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.15)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.15)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded -0.68)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.26)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.84)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 690
\-------------------------

Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2521; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.58)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.75)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.33)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.23)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.23)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.23)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.23)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.58)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.31)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.90)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.90)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.90)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.90)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.02)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.02)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.44)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.69)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.97)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.14)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.50)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.09)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.84)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.84)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.84)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.84)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.61)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded -0.86)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 691
\-------------------------

Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2516; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.97)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 1.13)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.48)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.48)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.48)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.48)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.90)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.98)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.38)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.62)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.16)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'right')
Agent followed the waypoint right. (rewarded 2.40)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 692
\-------------------------

Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2511; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.85)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.96)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.89)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.93)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.93)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.38)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.06)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.06)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.38)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.01)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.01)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.01)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.01)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.85)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.85)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.03)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.35)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.39)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.07)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.00)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.72)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.60)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.00)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'right')
Agent followed the waypoint right. (rewarded 1.90)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.45)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.03)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.03)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.03)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.17)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 0.47)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 0.47)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 0.47)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.28)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.14)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.02)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.21)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 693
\-------------------------

Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2506; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.80)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.44)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.78)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.27)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.77)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.77)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.72)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.72)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.02)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.02)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.02)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.02)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.02)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 1.53)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.36)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 0.76)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.90)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent drove right instead of forward. (rewarded 1.07)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.77)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded -0.25)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.70)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 694
\-------------------------

Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2501; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.65)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.18)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.42)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.83)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.91)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.85)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.32)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.14)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.13)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.94)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 695
\-------------------------

Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2496; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.72)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.53)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.30)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.21)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.19)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.13)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.77)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.30)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 696
\-------------------------

Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2491; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.35)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'right')
Agent drove right instead of forward. (rewarded 0.29)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.84)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.84)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.84)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.15)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.06)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.57)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.81)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.72)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 0.97)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.25)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.49)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.49)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.63)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.63)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.10)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.10)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.10)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.69)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.30)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.30)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.30)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.30)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.07)
33% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 697
\-------------------------

Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2486; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove left instead of forward. (rewarded 1.10)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.45)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.29)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.29)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.26)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.26)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.77)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.77)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.79)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 698
\-------------------------

Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2481; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.03)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.39)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.76)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.96)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.96)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.96)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.96)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.96)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.96)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.15)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.23)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.24)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.24)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.24)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.24)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.24)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.64)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 699
\-------------------------

Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2476; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove right instead of forward. (rewarded 0.87)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.24)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.26)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 1.83)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.60)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.15)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.43)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.60)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.85)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.08)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.21)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.73)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 700
\-------------------------

Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2471; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.17)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.73)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.06)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.87)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.31)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.62)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.82)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.94)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 701
\-------------------------

Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2466; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.56)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.68)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.65)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.26)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.26)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.26)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.72)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.81)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.90)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.46)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.46)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.46)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.46)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.46)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.46)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.96)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.47)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.47)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.24)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.59)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded 0.20)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.57)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.94)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.55)
16% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 702
\-------------------------

Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2461; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 0.38)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.72)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.04)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.25)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.25)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.25)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.54)
86% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.54)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.56)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.76)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.04)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.55)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.55)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.55)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.55)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.55)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.55)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.55)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.35)
71% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 703
\-------------------------

Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2456; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.36)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.04)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.91)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.59)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.59)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.59)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.59)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.59)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.59)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.62)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.62)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.62)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.62)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.69)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.73)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.33)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.28)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.31)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 704
\-------------------------

Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2451; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.41)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.52)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 0.94)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.93)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.93)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.79)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.41)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.98)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.45)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.48)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.16)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.16)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.16)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.27)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.73)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.45)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.51)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.40)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.29)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.18)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.18)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.18)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.18)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.18)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.60)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 1.23)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 0.70)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 0.70)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.54)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.76)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 705
\-------------------------

Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2446; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.92)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.36)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.05)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.25)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.75)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.75)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.75)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.75)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.80)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.80)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.04)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.95)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.28)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.28)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.63)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.70)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.70)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.70)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.05)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.08)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.08)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.29)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.24)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.02)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.10)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.27)
33% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 706
\-------------------------

Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2441; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.34)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.98)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.98)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.98)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.98)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.98)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.98)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.78)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 1.12)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.83)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.19)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 1.30)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.28)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.28)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.28)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.50)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.50)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.34)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.92)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.68)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.22)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 1.10)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.04)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.89)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.25)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.72)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.72)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.72)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.22)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.83)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.86)
4% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 707
\-------------------------

Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2437; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent drove forward instead of right. (rewarded 1.76)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.42)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.29)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.06)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.06)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.37)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.89)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 708
\-------------------------

Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2432; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.08)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.22)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.46)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.46)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.46)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.46)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.46)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.46)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.57)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.66)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.66)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 1.62)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.66)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.28)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.51)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 709
\-------------------------

Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2427; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.76)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.15)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.88)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.58)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.58)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.58)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.58)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.58)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.75)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.69)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.69)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.69)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.66)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.44)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.44)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.44)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.53)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.09)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.09)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 0.90)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.85)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.85)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.85)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.54)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.61)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 1.41)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.53)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.91)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.94)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.94)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.94)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.94)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.94)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.94)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.57)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.55)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 710
\-------------------------

Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2422; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.94)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.75)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.69)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.68)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.93)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 1.32)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.36)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.76)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.07)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.07)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.07)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.12)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.96)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.44)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.43)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.50)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.67)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.28)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.28)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.28)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.28)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.28)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.28)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.74)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 711
\-------------------------

Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2417; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.35)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.60)
93% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.60)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.06)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.06)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.06)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.06)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.06)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.00)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.15)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.63)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 712
\-------------------------

Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2412; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.54)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.76)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.46)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.46)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.46)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.46)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.46)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.46)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.52)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.53)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.47)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.47)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.47)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.33)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.23)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 713
\-------------------------

Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2407; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.64)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.57)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.97)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.15)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.15)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.15)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 2.80)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.48)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.48)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.48)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.10)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.86)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.32)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.32)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.22)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.22)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.16)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.14)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.32)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.61)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.23)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.23)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.23)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.23)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.01)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.45)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.43)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.44)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.44)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.05)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 714
\-------------------------

Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2403; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.08)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.45)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.27)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.02)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.26)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.36)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.93)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.93)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.93)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.43)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.92)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded -0.10)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.93)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.21)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.71)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.68)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.25)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.25)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.25)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.54)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.55)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.37)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.16)
16% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 715
\-------------------------

Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2398; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.98)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.42)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.35)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.35)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.74)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.88)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.76)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.76)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.76)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.76)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.24)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.99)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.77)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.86)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.74)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.33)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.92)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.40)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.29)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.29)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.29)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.29)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.78)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.59)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.84)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 716
\-------------------------

Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2393; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.04)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.82)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.38)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.38)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.38)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.31)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.79)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.94)
72% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 717
\-------------------------

Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2388; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.30)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.18)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.62)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.26)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.24)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.80)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.44)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.44)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.46)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.46)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.46)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.46)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.46)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.64)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.16)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.76)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.13)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.13)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.13)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.20)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.57)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.03)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.88)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.81)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 718
\-------------------------

Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2384; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.06)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.68)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.51)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.74)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.27)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.92)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.84)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.84)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.84)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.84)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.06)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.44)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.66)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.66)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.08)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.08)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.08)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.86)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.75)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
43% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 719
\-------------------------

Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2379; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'left')
Agent drove forward instead of left. (rewarded 0.70)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.27)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.58)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.17)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.07)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.38)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.38)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.38)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.59)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.57)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.57)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.57)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.60)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 720
\-------------------------

Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2374; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.85)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.64)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.04)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.04)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.04)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.04)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.14)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.15)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 721
\-------------------------

Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2369; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.38)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.27)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.19)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.68)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.03)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.70)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.70)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.51)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.51)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.12)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.12)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.15)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.52)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.89)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.42)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.83)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.88)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 1.48)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.49)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.18)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.73)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 722
\-------------------------

Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2365; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.32)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.81)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.60)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.89)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.43)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.99)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.20)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.67)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.82)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.36)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 723
\-------------------------

Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2360; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.44)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.10)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.11)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.76)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.46)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.28)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.59)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.60)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.21)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.01)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.17)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.72)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.93)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.43)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.02)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.73)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.73)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.03)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.97)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
37% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
37% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
37% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
37% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.81)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.36)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.94)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.45)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.26)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.42)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded -0.18)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.27)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 0.26)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.24)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.37)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 724
\-------------------------

Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2355; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.50)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.95)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.59)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.59)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.59)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.10)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.11)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.58)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.03)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.39)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.05)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.50)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.96)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.55)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.94)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.94)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.62)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.64)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 725
\-------------------------

Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2350; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.37)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.78)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.01)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent drove forward instead of right. (rewarded 1.25)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.46)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 726
\-------------------------

Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2346; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.33)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.45)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.28)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.28)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.28)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.28)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.28)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.28)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.09)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.99)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.92)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.25)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.25)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.25)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.25)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.25)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.69)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.60)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.60)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.25)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.86)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.33)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.33)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.48)
36% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 727
\-------------------------

Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2341; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.01)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.01)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.01)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.01)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.62)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.62)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.10)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.93)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.93)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.93)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.13)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.13)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.13)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.82)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.00)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.34)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.24)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 728
\-------------------------

Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2336; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.78)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.81)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.01)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.12)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.47)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.27)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.82)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 0.30)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.11)
63% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.11)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.94)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.94)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.94)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 2.70)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 2.70)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 2.70)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.57)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.43)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.89)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.09)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.09)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.09)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.09)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.03)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.92)
37% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 729
\-------------------------

Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2332; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.24)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 2.15)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 1.50)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.02)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.30)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.30)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.30)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.07)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.07)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.07)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.07)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.36)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.20)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.10)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 730
\-------------------------

Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2327; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.98)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.35)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.08)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.38)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.85)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.85)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.85)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.85)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.50)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.99)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.18)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.83)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.81)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.12)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.44)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.27)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.59)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.27)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 731
\-------------------------

Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2322; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.97)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.07)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.23)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.23)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.20)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.29)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.04)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.01)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.01)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.01)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.47)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.97)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.97)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.97)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.97)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.73)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 1.16)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.12)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.12)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.52)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.52)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.82)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.66)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 732
\-------------------------

Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2318; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.76)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.50)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.94)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.00)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.63)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.79)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.40)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.93)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.93)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.93)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.93)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.55)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.06)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.83)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.03)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.70)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.56)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.33)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.77)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.06)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.13)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.26)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 0.36)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 733
\-------------------------

Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2313; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.96)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.23)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.83)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.75)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.23)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.56)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.95)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.95)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.95)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.95)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.05)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.37)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.37)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.37)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.37)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.64)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.67)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.04)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.13)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.13)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.71)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.71)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.71)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.11)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 734
\-------------------------

Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2308; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.98)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.66)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.59)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.78)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.36)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.81)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.39)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.39)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.39)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.39)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.39)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.39)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.39)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.10)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.12)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.82)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.10)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.10)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.10)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.85)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.95)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.93)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.93)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.93)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.93)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.95)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 735
\-------------------------

Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2304; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.09)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.64)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.09)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.92)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.27)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 1.80)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.79)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.60)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.60)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.60)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.94)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.95)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.95)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded -0.03)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.26)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.97)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.29)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.29)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.29)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.29)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.29)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.29)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.29)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.29)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.11)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.37)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.21)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.60)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.12)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.12)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.12)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.46)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 736
\-------------------------

Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2299; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.54)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.24)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.25)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.80)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.02)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.97)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.42)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.12)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.04)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 737
\-------------------------

Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2295; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.05)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.73)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.21)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.21)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.21)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.21)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.21)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.21)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.84)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.18)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.19)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 738
\-------------------------

Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2290; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.12)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.49)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.53)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.53)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.53)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.53)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.75)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.78)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.19)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.15)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 739
\-------------------------

Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2286; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.66)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.79)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.85)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.47)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.13)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.03)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.33)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.39)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.39)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.39)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.39)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.37)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.94)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.62)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.95)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.44)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.47)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.28)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.67)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.74)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 740
\-------------------------

Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2281; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.47)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.51)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.51)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.51)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.51)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.51)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.39)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.29)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.29)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.14)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.79)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.89)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.52)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.52)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.52)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.52)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.62)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.38)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.76)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.20)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.20)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.20)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.20)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.65)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.48)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.80)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 741
\-------------------------

Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2276; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.34)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 0.92)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.65)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.16)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.17)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.83)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.83)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.88)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.88)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.88)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.88)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.89)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.46)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.41)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.38)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.74)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.81)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.04)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.04)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.04)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.53)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.69)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.69)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.69)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 742
\-------------------------

Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2272; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.76)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.76)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.78)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.78)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.78)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.13)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.16)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.16)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.44)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.10)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.10)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.10)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.10)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.94)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.76)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 743
\-------------------------

Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2267; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.99)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.30)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.91)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.64)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.17)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.42)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.42)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.42)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.42)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.42)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.94)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.46)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.46)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.51)
63% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.51)
63% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.51)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.27)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.37)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.12)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.89)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.89)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.89)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.89)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.58)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded -0.27)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded -0.27)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.60)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.60)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.60)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.45)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.42)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.24)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.18)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.18)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.18)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.18)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.27)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.58)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 0.56)
7% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 0.56)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.42)
3% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 744
\-------------------------

Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2263; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.03)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.29)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.74)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.74)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.74)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.74)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.71)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.28)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded 0.20)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.75)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.48)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.22)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.22)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.22)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.22)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.45)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.23)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.23)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.31)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.35)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.16)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.16)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.16)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.07)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.07)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.07)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.95)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.64)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.64)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.64)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.41)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.97)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.70)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.67)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.39)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.60)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 745
\-------------------------

Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2258; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.98)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.46)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.04)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.04)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.04)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.04)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.04)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.83)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.83)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.83)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.50)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.12)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.31)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.06)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.06)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.87)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.90)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.00)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.29)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.28)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 746
\-------------------------

Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2254; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.13)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.63)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.81)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.35)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.18)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.35)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.47)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.47)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.05)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded -0.08)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.00)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.18)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.19)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.04)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.06)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 747
\-------------------------

Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2249; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.94)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.37)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.83)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.83)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.83)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.83)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.19)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.26)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.02)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.02)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.02)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.48)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.29)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.63)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.08)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.51)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 748
\-------------------------

Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2245; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.17)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.06)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.53)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.48)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.57)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.23)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.86)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.57)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.93)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.93)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.93)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.93)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.93)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.24)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.15)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.89)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.89)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.89)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.59)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.59)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.55)
36% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 749
\-------------------------

Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2240; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.84)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.30)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.99)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.77)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.99)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.99)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.99)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.99)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.99)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.99)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.99)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.82)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.46)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.38)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 750
\-------------------------

Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2236; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.29)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.50)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.55)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.63)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 751
\-------------------------

Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2231; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.97)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.31)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.47)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.47)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.47)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.54)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.69)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.72)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.21)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.21)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.90)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.83)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.25)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.73)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 752
\-------------------------

Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2227; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.09)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.48)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.84)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.27)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.23)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.51)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.40)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.40)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.78)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.47)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.49)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent followed the waypoint forward. (rewarded 1.50)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.93)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 753
\-------------------------

Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2222; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.75)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.48)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.07)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.28)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.28)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.28)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.28)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.35)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.36)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.36)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.36)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.36)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.34)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.46)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.46)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.91)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.56)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.16)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.16)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.16)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.16)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.52)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.46)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.46)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.46)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.46)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.46)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.09)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.03)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 754
\-------------------------

Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2218; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.77)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.81)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.28)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.11)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.08)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.67)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.67)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.80)
72% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 755
\-------------------------

Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2214; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.75)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.49)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.49)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.02)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.02)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.02)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.02)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.13)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.71)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.71)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.71)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.71)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.72)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.89)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded -0.17)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.60)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 756
\-------------------------

Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2209; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.24)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.89)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.42)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.73)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.15)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.37)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.73)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.60)
69% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.60)
69% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.60)
69% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.60)
69% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.60)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.34)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.20)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.31)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 757
\-------------------------

Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2205; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.81)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.71)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.73)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.78)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.83)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.80)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.20)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.85)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.15)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.76)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.61)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.59)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.19)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.07)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.78)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.61)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.09)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 758
\-------------------------

Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2200; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 1.35)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.72)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.55)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.82)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.50)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.95)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.10)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.41)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 759
\-------------------------

Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2196; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.16)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.85)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.21)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.21)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.87)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.87)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.31)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.89)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.79)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.79)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.79)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.79)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.07)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.72)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.97)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', 'right')
Agent drove forward instead of right. (rewarded 0.74)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.77)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.49)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.15)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.15)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 760
\-------------------------

Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2191; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.73)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.75)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.75)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.75)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.75)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.39)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.32)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.32)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.32)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.54)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.43)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.23)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.40)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.28)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.27)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.37)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.94)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.43)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.43)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.43)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.43)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.11)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.08)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.95)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.56)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 761
\-------------------------

Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2187; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.10)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.65)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.35)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.77)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.77)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.00)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.51)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.12)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.44)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.44)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.25)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.03)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.45)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.45)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.26)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.81)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.21)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.68)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.94)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.71)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.71)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.71)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.71)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.50)
17% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 762
\-------------------------

Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2183; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.00)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.66)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.60)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.60)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.60)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.60)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.81)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.86)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.86)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.86)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.71)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.45)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.05)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 763
\-------------------------

Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2178; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.24)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.59)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.13)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.34)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.43)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.43)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.43)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.43)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.89)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 764
\-------------------------

Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2174; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.28)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.42)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.76)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.02)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.18)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.73)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.52)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.11)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.79)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.21)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.21)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.21)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.21)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.21)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.07)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.70)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.28)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.06)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.13)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.92)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.93)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 0.50)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.29)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.64)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 765
\-------------------------

Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2170; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.53)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.12)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.95)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.52)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.61)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.79)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 766
\-------------------------

Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2165; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded 1.20)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.86)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.51)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.08)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.08)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.08)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.98)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.40)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.50)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.50)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.66)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.66)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.37)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.74)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.52)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded -0.14)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.75)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 0.62)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.66)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.66)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.66)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.66)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.75)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.73)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.95)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.71)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.80)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 767
\-------------------------

Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2161; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.41)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.07)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.62)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.35)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.35)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.35)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.35)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.85)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.08)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.08)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.63)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.81)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.11)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.15)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.70)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 768
\-------------------------

Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2157; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.17)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.15)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.93)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.48)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.48)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.48)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.48)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.61)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.00)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.55)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.85)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.85)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.85)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.85)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.82)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.91)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.02)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.88)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 769
\-------------------------

Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2152; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.12)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.30)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.53)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.70)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.58)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.58)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.59)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.38)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.38)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.38)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.14)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.19)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.91)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 770
\-------------------------

Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2148; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.04)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.79)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.61)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.69)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.01)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.01)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.57)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.57)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.71)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.85)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.15)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 771
\-------------------------

Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2144; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.64)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.22)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.41)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.15)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.60)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.98)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.98)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.98)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.98)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.68)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.07)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.37)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.34)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.34)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.80)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.65)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.29)
43% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 772
\-------------------------

Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2140; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.57)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.53)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', None)
Agent drove forward instead of right. (rewarded 1.51)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 0.97)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.84)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.44)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.31)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.31)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.31)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.87)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.53)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.79)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.79)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.79)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.79)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.12)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.47)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.47)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.34)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.34)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.34)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.34)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.36)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.24)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.84)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 773
\-------------------------

Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2135; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.65)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.81)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.46)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.43)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.63)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.83)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.17)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.26)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.54)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.41)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.76)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.59)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 0.60)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.01)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.09)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 774
\-------------------------

Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2131; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.93)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.45)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.92)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.50)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.95)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.95)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.95)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.44)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.41)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 775
\-------------------------

Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2127; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.61)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.27)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.19)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 776
\-------------------------

Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2122; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.49)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.07)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.33)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.10)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.18)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.18)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.18)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.18)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.38)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.38)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.38)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.38)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.38)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.86)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.52)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.52)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.52)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.52)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.29)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.44)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.71)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.05)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.44)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.44)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.44)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.69)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.09)
47% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 777
\-------------------------

Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2118; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.44)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.59)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.39)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.43)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.43)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.78)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 1.89)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.66)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.34)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.34)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.34)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.34)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.34)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.11)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.45)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.26)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.70)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.55)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.26)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.47)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.16)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 0.79)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 778
\-------------------------

Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2114; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.92)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.61)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.18)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.38)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.38)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.38)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.50)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.82)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.77)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.08)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.30)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.34)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.04)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.23)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.10)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.65)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 779
\-------------------------

Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2110; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.28)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.97)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.33)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.33)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.03)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.05)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.83)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.50)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.99)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.99)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.72)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.52)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.52)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.96)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 780
\-------------------------

Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2106; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.57)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.99)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.23)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.23)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.75)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.75)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.75)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.75)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.27)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.44)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.89)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.46)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.46)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.46)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.30)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.67)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'right')
Agent properly idled at a red light. (rewarded 0.90)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.66)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.65)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 781
\-------------------------

Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2101; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.51)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.38)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.34)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.21)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.28)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.41)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.15)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.79)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 782
\-------------------------

Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2097; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.43)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.31)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.40)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.33)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.33)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.33)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.50)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.23)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.23)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.23)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 1.20)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.21)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.06)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.63)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.44)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.13)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.27)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.27)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.27)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.27)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.93)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.99)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 783
\-------------------------

Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2093; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.65)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.23)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.92)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.02)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.82)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.17)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.67)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.14)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.48)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.48)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.48)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.48)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.24)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.83)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 784
\-------------------------

Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2089; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.55)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.29)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.89)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.89)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.99)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.76)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.80)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.00)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.16)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.70)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.67)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.83)
57% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 785
\-------------------------

Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2085; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.99)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.19)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.29)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.16)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.18)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.19)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.13)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.54)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.54)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.95)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.96)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.96)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded -0.20)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.68)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.81)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.14)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.54)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.06)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.04)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.41)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.59)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.59)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.59)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.59)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.85)
4% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 786
\-------------------------

Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2080; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.43)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.95)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.73)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.72)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.99)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.49)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.97)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.99)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 1.19)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.05)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.81)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.81)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.81)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.81)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.81)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.11)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded -0.21)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.27)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.27)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.53)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.49)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.35)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.69)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 1.38)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 787
\-------------------------

Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2076; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.39)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.70)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.96)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.96)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.96)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.96)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.23)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.86)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.75)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.76)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.76)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.76)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.76)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.20)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.05)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 788
\-------------------------

Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2072; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.47)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.75)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.18)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.18)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.18)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.18)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.35)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.01)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.40)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.40)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.40)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.68)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.40)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.04)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.64)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.00)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.70)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.69)
43% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 789
\-------------------------

Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2068; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.41)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.51)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.03)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.59)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.95)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.36)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.36)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.36)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.36)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.36)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.36)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.63)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.88)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.64)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.97)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.26)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.00)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.40)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.40)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.40)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.16)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 790
\-------------------------

Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2064; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.22)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.22)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.47)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.47)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.47)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.47)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.47)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.47)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.47)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.05)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.04)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.49)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.65)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.20)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 791
\-------------------------

Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2060; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.54)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.90)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.27)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.89)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.89)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.89)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.15)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.15)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.32)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 792
\-------------------------

Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2056; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.03)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.99)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.08)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.30)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.87)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.41)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.23)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.11)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded -0.11)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded -0.11)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded -0.11)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded -0.11)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.99)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 793
\-------------------------

Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2052; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.13)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.88)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.30)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.74)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.52)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.10)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.10)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.10)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.11)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.76)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.66)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.66)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.53)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.53)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.53)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.53)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.75)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.18)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.44)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.56)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.75)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.29)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.18)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.27)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 794
\-------------------------

Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2047; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.86)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.48)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.19)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.14)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.33)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.53)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 795
\-------------------------

Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2043; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.01)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.59)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.02)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.63)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.44)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.93)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.94)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.90)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.62)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.62)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.90)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.44)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 1.55)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 1.55)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 1.55)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove forward instead of left. (rewarded 0.52)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove forward instead of left. (rewarded 0.52)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded -0.12)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.21)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.21)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.21)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.06)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent drove right instead of forward. (rewarded -0.23)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 796
\-------------------------

Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2039; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.52)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.47)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.49)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.76)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.55)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.39)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.73)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.62)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.57)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 797
\-------------------------

Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2035; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.91)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.06)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.95)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.15)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.15)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.15)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.15)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.15)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.40)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.94)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.49)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.49)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.49)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.25)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.49)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 798
\-------------------------

Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2031; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.54)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.39)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.93)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.20)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 799
\-------------------------

Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2027; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.26)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.26)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.55)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.38)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.94)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.94)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.71)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 0.91)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 800
\-------------------------

Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2023; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.77)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.23)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.93)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.82)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.82)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.82)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.01)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded -0.07)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.78)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.78)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.78)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.78)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove forward instead of right. (rewarded 1.47)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.72)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.72)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.72)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.72)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.63)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded -0.24)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.94)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.94)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.94)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.94)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.44)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.03)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.76)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded -0.55)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded -0.55)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded -0.55)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.19)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.10)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 801
\-------------------------

Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2019; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.55)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.11)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.54)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.27)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.55)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.82)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.95)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.95)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.76)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.16)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.05)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.68)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.43)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.43)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.58)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.38)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.66)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.00)
33% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 802
\-------------------------

Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2015; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.02)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.59)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.45)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.05)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.52)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.43)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.44)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.44)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 803
\-------------------------

Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2011; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.48)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.03)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.25)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.07)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.62)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.30)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.00)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.48)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.48)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.48)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.48)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.62)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.95)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.77)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.17)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.72)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.09)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.09)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.45)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.11)
49% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 804
\-------------------------

Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2007; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.39)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.61)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 2.78)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 2.78)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 2.78)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 1.55)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.21)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.29)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.72)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.13)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.25)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.02)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.69)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.82)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.18)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.51)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.10)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.89)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.55)
37% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 805
\-------------------------

Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.2003; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.83)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.92)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 1.04)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.27)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 1.16)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 1.50)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.29)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.29)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.55)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.55)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.55)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.55)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.08)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.48)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.48)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.48)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.48)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.27)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 806
\-------------------------

Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1999; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.77)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.15)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.46)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.02)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.02)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.58)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.10)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.10)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.10)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.55)
72% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 807
\-------------------------

Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1995; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.62)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.51)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.04)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.73)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.73)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.73)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.86)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded -0.08)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded -0.08)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded -0.08)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded -0.08)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.61)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.41)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.04)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.50)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 808
\-------------------------

Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1991; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.02)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.53)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.43)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.43)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.42)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.74)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.74)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.92)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded -0.03)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.73)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 809
\-------------------------

Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1987; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.94)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.34)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.34)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.34)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.34)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.34)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.09)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.36)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.44)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.66)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.86)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.20)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.20)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.20)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.20)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.20)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.87)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.49)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.51)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 0.65)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.58)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.58)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.58)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.58)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.63)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 810
\-------------------------

Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1983; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 2.31)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.21)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.95)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.89)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 1.37)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.17)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 1.24)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.89)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.73)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.75)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.75)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.75)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.58)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.58)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.10)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.59)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.10)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.76)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.51)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.63)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.59)
17% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.59)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.64)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.36)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.36)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.36)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.36)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.36)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.75)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.14)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.70)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 811
\-------------------------

Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1979; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.82)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.69)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.85)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.24)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.37)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.56)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.13)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 812
\-------------------------

Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1975; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.51)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 2.62)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.75)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.28)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.28)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.28)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.02)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.75)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.75)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.75)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'right')
Agent drove forward instead of left. (rewarded 1.70)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.76)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.74)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.51)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.52)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 813
\-------------------------

Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1971; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.36)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.29)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.55)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.47)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.76)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.13)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.72)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.74)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.47)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.32)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'right')
Agent properly idled at a red light. (rewarded 2.08)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'right')
Agent properly idled at a red light. (rewarded 2.08)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'right')
Agent properly idled at a red light. (rewarded 2.08)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'right')
Agent properly idled at a red light. (rewarded 2.08)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.39)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.38)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.13)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.13)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.53)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.39)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent drove forward instead of right. (rewarded 0.43)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 814
\-------------------------

Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1967; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.11)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.18)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.88)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.64)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.22)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.09)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 815
\-------------------------

Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1963; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.13)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.21)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.17)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.88)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.55)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.27)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 816
\-------------------------

Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1959; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.59)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.98)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.39)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.68)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.68)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.68)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.72)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.86)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.86)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.86)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.86)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.86)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'right')
Agent followed the waypoint forward. (rewarded 2.23)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.70)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.02)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.58)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.99)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.70)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.70)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.70)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.46)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.44)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.46)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 817
\-------------------------

Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1955; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.58)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.64)
92% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.64)
92% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.64)
92% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.64)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.36)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.27)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.27)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.70)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 818
\-------------------------

Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1951; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.88)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.01)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.42)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.70)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.77)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.65)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.38)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.38)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.38)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.38)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.61)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.08)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.06)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 0.05)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 0.05)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.43)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.43)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.35)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 819
\-------------------------

Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1948; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.41)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.80)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.07)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.39)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.39)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.82)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 820
\-------------------------

Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1944; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.97)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.86)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.65)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.63)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.18)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.18)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.18)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.46)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.52)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.19)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.19)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.19)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.54)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 821
\-------------------------

Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1940; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.03)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.91)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.19)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.30)
76% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 822
\-------------------------

Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1936; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.14)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.38)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.54)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.52)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.52)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.14)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.58)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.58)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.58)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.58)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.29)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.21)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove right instead of forward. (rewarded 0.26)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.04)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.84)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.36)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.21)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.21)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.21)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.59)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.43)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.43)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.43)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.43)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.13)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.80)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.06)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.45)
4% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 823
\-------------------------

Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1932; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 2.79)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.31)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.18)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.05)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.94)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.94)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.94)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.76)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.49)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.49)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.49)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.49)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.47)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.42)
67% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 824
\-------------------------

Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1928; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.37)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.55)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.58)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.57)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.04)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.18)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.18)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.43)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.43)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.43)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.43)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.95)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.62)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.34)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.14)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'right')
Agent drove right instead of forward. (rewarded -0.16)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.41)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.87)
24% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 825
\-------------------------

Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1924; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.11)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.26)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.85)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.85)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.00)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.00)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.00)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.37)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.82)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.46)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.49)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent followed the waypoint right. (rewarded 2.60)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.33)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.11)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 826
\-------------------------

Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1920; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.87)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.08)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.07)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.07)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.07)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.07)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.07)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.07)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.78)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.74)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.26)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.26)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.54)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.79)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.98)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.47)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 827
\-------------------------

Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1917; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.69)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.09)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.80)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.63)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.94)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.11)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 828
\-------------------------

Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1913; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.97)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.76)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.61)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.61)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.53)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.42)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.42)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.02)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.02)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.02)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.76)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.63)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 0.56)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.13)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 829
\-------------------------

Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1909; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.76)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.30)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.93)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.93)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.93)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.93)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.47)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.30)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.87)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.04)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.07)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.03)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.73)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.73)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.73)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.73)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.77)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.08)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.08)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.25)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.21)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.77)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.07)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 830
\-------------------------

Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1905; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.21)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.79)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.90)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.52)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.83)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.83)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.50)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.54)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.82)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.52)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.08)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.08)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.08)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.08)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.62)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.23)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.97)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 831
\-------------------------

Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1901; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.04)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.63)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.70)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.77)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 1.49)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.14)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.47)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.14)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 0.94)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.64)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 832
\-------------------------

Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1898; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.62)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.34)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.38)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.57)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.82)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.49)
76% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 833
\-------------------------

Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1894; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.99)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.77)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.04)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.04)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.04)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.04)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.04)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.04)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.91)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.40)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.29)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.29)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.58)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.58)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.58)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.07)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.44)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.44)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'forward')
Agent properly idled at a red light. (rewarded 2.23)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.76)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.63)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.63)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.63)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.63)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.63)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.14)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.72)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 834
\-------------------------

Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1890; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.11)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.51)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded 0.68)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.56)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 1.70)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.52)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.88)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.80)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.78)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.16)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.99)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.89)
66% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 835
\-------------------------

Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1886; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent drove right instead of left. (rewarded 1.47)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.48)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.34)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.90)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.31)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.16)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.48)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.61)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.01)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.01)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.01)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.01)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.01)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.01)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.34)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 1.23)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 836
\-------------------------

Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1882; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.80)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.21)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.53)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.88)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.43)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.43)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.34)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.46)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.31)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.31)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.65)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.65)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.65)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.65)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.65)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.65)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.65)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.18)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.26)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.71)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.71)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.71)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.71)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.46)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded -0.12)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 0.79)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.51)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 837
\-------------------------

Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1879; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.12)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.44)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.44)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.44)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.44)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.44)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.44)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.36)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.70)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.09)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.09)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.09)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.32)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.61)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.37)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.12)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 2.65)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.18)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.93)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 838
\-------------------------

Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1875; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.35)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.75)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.08)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.08)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.08)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.08)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.08)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.08)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.08)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.08)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.47)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.61)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.61)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.61)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.61)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.54)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.63)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.84)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.07)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.29)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.23)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.52)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.64)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 839
\-------------------------

Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1871; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.93)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.48)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.35)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.35)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.35)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.35)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.07)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.93)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.48)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 840
\-------------------------

Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1867; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.42)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.39)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.89)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.82)
84% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.82)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.89)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.61)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.61)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.62)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.82)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.16)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.49)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.11)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.45)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.45)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.50)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.64)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.64)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.64)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.64)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.32)
36% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 841
\-------------------------

Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1864; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.94)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.72)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.77)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.76)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.76)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.76)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.76)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.95)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.95)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.67)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.44)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.29)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.30)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.30)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.96)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.96)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.96)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.96)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.24)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.24)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.20)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
43% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 842
\-------------------------

Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1860; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.46)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.70)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.45)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.45)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.45)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.90)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.82)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.53)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.53)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.34)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.34)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.42)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.42)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.42)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.42)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.59)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.12)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.20)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.62)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.94)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.41)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.81)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.10)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.10)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.10)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.10)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.10)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.07)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 843
\-------------------------

Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1856; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.41)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.51)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.07)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.93)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.93)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.93)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.93)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.46)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.30)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.30)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.30)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.30)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.37)
72% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 844
\-------------------------

Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1853; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.15)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.21)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.94)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.16)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.71)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.53)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.14)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.53)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.53)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.32)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.20)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.98)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.98)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.98)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.08)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.48)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.46)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.46)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.62)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.18)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 0.41)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 845
\-------------------------

Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1849; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.48)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.96)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.40)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.05)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.09)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.09)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.09)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.09)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.64)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.27)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.06)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.81)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.85)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 2.28)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.80)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.23)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 846
\-------------------------

Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1845; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.79)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.89)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.32)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.76)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.76)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.76)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.10)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.48)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.81)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.81)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.81)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.81)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.31)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.62)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.62)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.62)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.62)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.62)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.08)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.67)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.40)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.60)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.93)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
43% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 847
\-------------------------

Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1842; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.82)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.85)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.49)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.49)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.12)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.67)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.56)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.75)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.20)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 848
\-------------------------

Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1838; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.57)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.23)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded -0.00)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.34)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.50)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.50)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.50)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.50)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.33)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded -0.04)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded -0.04)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.78)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.74)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.74)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.23)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.23)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.23)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.23)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.92)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.46)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.96)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.96)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.96)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.19)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.89)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.52)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.89)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.48)
5% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.48)
5% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.48)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.61)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 849
\-------------------------

Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1834; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.98)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.34)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.29)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.29)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.29)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.29)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.29)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.29)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.29)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.29)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.29)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.29)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.72)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.72)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.72)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.83)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.31)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.00)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.64)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.54)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.54)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.57)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.54)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 850
\-------------------------

Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1830; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.41)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.84)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.23)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.23)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.23)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.09)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.15)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.15)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.15)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.80)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 851
\-------------------------

Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1827; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.75)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.62)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.55)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.58)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.07)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.45)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.99)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 852
\-------------------------

Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1823; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.93)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.88)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.64)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.76)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.17)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.85)
72% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 853
\-------------------------

Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1820; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.93)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.08)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.68)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.68)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.57)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.51)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.01)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 854
\-------------------------

Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1816; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.98)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.19)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.09)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.09)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.09)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 2.15)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.05)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.27)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.12)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 855
\-------------------------

Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1812; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.35)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.39)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.33)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 1.78)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.49)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.64)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.15)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.37)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.09)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.37)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 856
\-------------------------

Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1809; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent followed the waypoint right. (rewarded 2.03)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.80)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.41)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.26)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.26)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.26)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.70)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.77)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.77)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.77)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.77)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.77)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.77)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.48)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.32)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.57)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 857
\-------------------------

Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1805; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.65)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.06)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.87)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.31)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.44)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.43)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.75)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.75)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.75)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.75)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.41)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.41)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.55)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.09)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.14)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.14)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.14)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.14)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.14)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.81)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 858
\-------------------------

Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1801; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.10)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.40)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.12)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.31)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.72)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.93)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.84)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.22)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.11)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.11)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.11)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.85)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.51)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 859
\-------------------------

Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1798; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.73)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.76)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 2.30)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 2.30)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 2.30)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 2.30)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 2.30)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove left instead of forward. (rewarded 0.80)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.05)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.88)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.88)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.88)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.55)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.55)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.55)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.55)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.40)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.69)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.69)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.69)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.69)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.55)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.10)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent drove forward instead of right. (rewarded 1.68)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.14)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.03)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.20)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.81)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 860
\-------------------------

Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1794; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.49)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove right instead of left. (rewarded 1.62)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent drove left instead of forward. (rewarded 0.88)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.26)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.95)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.11)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.11)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 861
\-------------------------

Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1791; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.98)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.88)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.13)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.20)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.62)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.62)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.62)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.62)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.26)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.51)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.51)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.96)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.45)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.45)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.45)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.45)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.45)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.45)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.32)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.06)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.55)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.80)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.80)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.80)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.80)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.36)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.60)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 862
\-------------------------

Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1787; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.75)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.83)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.90)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.90)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.90)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.90)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.95)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.70)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove forward instead of left. (rewarded 0.34)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove forward instead of left. (rewarded 0.34)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove forward instead of left. (rewarded 0.34)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove forward instead of left. (rewarded 0.34)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.13)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.29)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.21)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.21)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.51)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 863
\-------------------------

Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1784; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.07)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.82)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.74)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.74)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.74)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.74)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.74)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.74)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.71)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.89)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.19)
76% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 864
\-------------------------

Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1780; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.87)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.12)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.64)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.48)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 865
\-------------------------

Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1776; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.76)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.88)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.14)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.14)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.14)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.14)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.14)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.90)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.79)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.02)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.38)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 0.24)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 0.24)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 0.24)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 0.24)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.84)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.54)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.29)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.54)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.54)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.54)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.54)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.32)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.18)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.09)
16% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 866
\-------------------------

Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1773; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.01)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.18)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.47)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.47)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.47)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.47)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.07)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.09)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.09)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.46)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.43)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.43)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.43)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.43)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.43)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.43)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.76)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded -0.06)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.64)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.44)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.55)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.55)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.55)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.29)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.15)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.56)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.15)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 867
\-------------------------

Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1769; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.13)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.21)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.39)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.31)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.37)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 868
\-------------------------

Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1766; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.19)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.47)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.06)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.06)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.38)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.91)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.57)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.57)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.57)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 2.88)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.89)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.54)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.54)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.54)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.67)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.12)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.55)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.55)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.55)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.55)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.55)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.99)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.58)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.58)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.58)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.58)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.58)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.38)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.52)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.68)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.35)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.68)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.56)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 869
\-------------------------

Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1762; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.08)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.75)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.09)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.20)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.82)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.43)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.26)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.47)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.47)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.47)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.47)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.94)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.91)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.55)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.99)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.34)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.34)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.73)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.12)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.93)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.68)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.80)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 870
\-------------------------

Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1759; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.42)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.95)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.58)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.51)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.51)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.51)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.51)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.63)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.73)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove forward instead of left. (rewarded 1.66)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove forward instead of left. (rewarded 1.66)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.97)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.91)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.88)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.14)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.32)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.20)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.20)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.20)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.20)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.81)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.15)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.41)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.41)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.41)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.41)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.55)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.83)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.83)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.35)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.35)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.35)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.13)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.13)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.13)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 871
\-------------------------

Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1755; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.99)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.05)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.79)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.79)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.78)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.35)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.50)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.21)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.72)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.72)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.72)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.72)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.76)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.57)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.62)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.93)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.93)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.54)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.92)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.01)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 872
\-------------------------

Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1752; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.73)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.16)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.91)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.91)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.91)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.91)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.91)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.91)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.82)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.14)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 1.46)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 873
\-------------------------

Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1748; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 3.00)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.27)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.68)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.69)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.97)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.68)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.68)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.68)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.29)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.34)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.47)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.10)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.13)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 874
\-------------------------

Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1745; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.37)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.45)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.86)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.08)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.46)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
77% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 875
\-------------------------

Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1741; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.05)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.08)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.96)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.53)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.53)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.52)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.68)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.01)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.49)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.58)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.86)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.20)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.63)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.20)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.29)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.95)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.29)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.58)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 876
\-------------------------

Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1738; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.31)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.22)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.09)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.00)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.14)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.62)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.42)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.71)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.50)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.50)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.37)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.37)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.37)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.37)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.36)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.19)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 877
\-------------------------

Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1734; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.03)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.21)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.46)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.89)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.41)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.41)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.41)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.41)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.41)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.42)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 878
\-------------------------

Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1731; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.85)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.25)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.14)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.71)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.82)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.29)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.29)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.29)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.29)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.59)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.21)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.10)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.10)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.49)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.01)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent drove left instead of right. (rewarded 1.29)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.64)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.51)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.51)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.21)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.21)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.18)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.18)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.83)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.02)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.02)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.02)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.02)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.26)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.43)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.17)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.06)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.06)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.37)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.37)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.37)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.80)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.32)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 879
\-------------------------

Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1727; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.41)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.54)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.12)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.84)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.33)
83% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.33)
83% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.33)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.66)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.56)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.57)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 0.99)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.04)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.81)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.73)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.41)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.48)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.39)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.42)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.42)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.49)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.98)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.89)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.83)
27% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.83)
27% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.83)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded -0.00)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded -0.00)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded -0.00)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.97)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.98)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.38)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.57)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.05)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.82)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 880
\-------------------------

Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1724; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 1.87)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.04)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.82)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded 1.75)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.62)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.43)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.43)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.43)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.43)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.43)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.48)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.83)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.07)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.60)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.60)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.60)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.60)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.45)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.65)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 1.22)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.36)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.98)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.82)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.77)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 0.78)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 881
\-------------------------

Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1720; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.58)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.39)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.76)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.55)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.42)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 2.24)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.16)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.16)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.16)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.16)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.16)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.34)
74% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 882
\-------------------------

Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1717; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.18)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.58)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.21)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.21)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.56)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.56)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.56)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.56)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.31)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.10)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.53)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.75)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.72)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.97)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.33)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 0.90)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'right')
Agent followed the waypoint forward. (rewarded 2.18)
57% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 883
\-------------------------

Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1714; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.79)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.11)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 2.87)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.36)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.31)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 884
\-------------------------

Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1710; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.33)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.81)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.57)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.60)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.47)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.89)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.72)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.64)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.64)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.64)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.64)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.64)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.90)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.71)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.48)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.08)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.08)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.45)
5% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 885
\-------------------------

Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1707; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.25)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.98)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.07)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.57)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.97)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.97)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.97)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.59)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.59)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.14)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 886
\-------------------------

Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1703; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.45)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.89)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.48)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.74)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.94)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.04)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.04)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.88)
73% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 887
\-------------------------

Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1700; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.45)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.23)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.23)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.23)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.23)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.23)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.18)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.18)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.09)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.75)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.98)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.98)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.98)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.98)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.28)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.53)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.32)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.01)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.29)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.92)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.92)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.92)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.92)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.89)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.79)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.29)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.29)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.29)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.29)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.72)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.24)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded -0.13)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.15)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.85)
4% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 888
\-------------------------

Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1697; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 2.25)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.27)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.19)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.73)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.16)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.60)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.62)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.96)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.60)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.60)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.60)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.27)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.27)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.27)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.27)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.10)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.05)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.05)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.05)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent followed the waypoint left. (rewarded 1.12)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.27)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.26)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.05)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.15)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.15)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.15)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.15)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.85)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.60)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.17)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 889
\-------------------------

Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1693; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.74)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.19)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.17)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.07)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.86)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.86)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.01)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.70)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.60)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.60)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.60)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.85)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.85)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.19)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.19)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.19)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.19)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.58)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.58)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.57)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.83)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 2.37)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.81)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.83)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.53)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.12)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.71)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.04)
17% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.04)
17% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.04)
17% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.04)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.90)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.28)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.94)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.23)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.23)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.07)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 890
\-------------------------

Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1690; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.11)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.42)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.69)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.67)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.79)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.31)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.09)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.73)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.73)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.73)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.42)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.23)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.28)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.67)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.67)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.67)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.67)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.67)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.67)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.57)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.80)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.80)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.80)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.80)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.80)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.80)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.41)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.12)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.04)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 891
\-------------------------

Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1686; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.90)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.41)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.96)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.96)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.35)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.51)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.63)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.60)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.60)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent followed the waypoint right. (rewarded 2.42)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.59)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.59)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.59)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.59)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.59)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.59)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.78)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.02)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.35)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.30)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.25)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 892
\-------------------------

Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1683; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.48)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.18)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.75)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.86)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.86)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.26)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.26)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.26)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.26)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.26)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.25)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.23)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.23)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.34)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.07)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.09)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 0.92)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.66)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 0.44)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 0.44)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.24)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.24)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.24)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.24)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.39)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.39)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.39)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.41)
27% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.41)
27% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.41)
27% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.41)
27% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.41)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.99)
23% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 893
\-------------------------

Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1680; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.79)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.00)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.38)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.31)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.17)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.12)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.12)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.12)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.12)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.12)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.01)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.83)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.45)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.62)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 894
\-------------------------

Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1676; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.26)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.06)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.61)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.71)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.91)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 895
\-------------------------

Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1673; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.83)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.04)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.63)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.46)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.22)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.97)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.98)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 896
\-------------------------

Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1670; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.77)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.79)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.83)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.99)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.58)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 897
\-------------------------

Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1666; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.48)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.06)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.81)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.53)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.53)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.53)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.53)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.53)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.41)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.07)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 898
\-------------------------

Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1663; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.76)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.16)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.88)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.78)
84% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.78)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.85)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.35)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.35)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.35)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.72)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.41)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.41)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.70)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.70)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.70)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.70)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.70)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.73)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.12)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.05)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.50)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.50)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.58)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 899
\-------------------------

Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1660; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.48)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.88)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.26)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.90)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 900
\-------------------------

Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1656; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.99)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.56)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.16)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.16)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.47)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.47)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.47)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.56)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 0.32)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.30)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.56)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 0.98)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 0.98)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.97)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.17)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 901
\-------------------------

Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1653; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.52)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.10)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.72)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.86)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.29)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.18)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.18)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.16)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.08)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.55)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.55)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.15)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.13)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.08)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.35)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.54)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.54)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.54)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.54)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.54)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.49)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.10)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.63)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.78)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 902
\-------------------------

Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1650; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.66)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.63)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.54)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 2.73)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.40)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.00)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.64)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.95)
73% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 903
\-------------------------

Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1646; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.12)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.22)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.82)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.44)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.99)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.32)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.75)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.85)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.06)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.06)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.46)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 904
\-------------------------

Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1643; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.06)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.27)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.25)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.25)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.25)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.25)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.25)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.47)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.15)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.15)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.15)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.08)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.14)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.73)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.18)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 905
\-------------------------

Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1640; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.79)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.52)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 0.72)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.94)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.44)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.44)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.44)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.69)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.77)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.77)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.01)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.81)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.81)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.81)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.43)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.11)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.11)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.11)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.21)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.49)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.39)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 906
\-------------------------

Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1637; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 1.51)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.59)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.96)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.99)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.78)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.71)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.40)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.46)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.78)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.78)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.78)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.28)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 907
\-------------------------

Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1633; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.18)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.02)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.47)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.47)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.47)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.47)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.47)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.67)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.28)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.28)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.28)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.28)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.43)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.88)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.88)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.88)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.88)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.88)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.88)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.53)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.62)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.74)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.88)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 0.91)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.16)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded -0.52)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded -0.52)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded -0.52)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded -0.52)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded -0.52)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded -0.52)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.43)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 908
\-------------------------

Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1630; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.05)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.87)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.83)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.53)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.37)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.52)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.75)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.43)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 0.56)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.22)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.22)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.22)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.22)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.22)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.08)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.39)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.16)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.57)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 909
\-------------------------

Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1627; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.07)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.97)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.72)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.11)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.93)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 910
\-------------------------

Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1624; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.23)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.22)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.45)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.45)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.45)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.45)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.45)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.36)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.47)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.54)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.01)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.17)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.10)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.10)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.69)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 1.76)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 1.76)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 1.76)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 1.76)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 1.76)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.17)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.10)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.15)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.22)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.71)
33% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 911
\-------------------------

Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1620; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.94)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.07)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.23)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.54)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.59)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.12)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.19)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.67)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.67)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.50)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.34)
57% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.34)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.02)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.26)
47% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 912
\-------------------------

Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1617; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.13)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.35)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.39)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.29)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.49)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.83)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 913
\-------------------------

Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1614; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.06)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.59)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.95)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.47)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.94)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.94)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.94)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.54)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.35)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.64)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.98)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.18)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.71)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.71)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.71)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.71)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.71)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.55)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.18)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.18)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.18)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.18)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.18)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 914
\-------------------------

Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1611; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.70)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.13)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.92)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.30)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.50)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.50)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.62)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.62)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.62)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.37)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.06)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.06)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.06)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.06)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.27)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.27)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.27)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.27)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.27)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.89)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.28)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.97)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.74)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.14)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.14)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.14)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.45)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 915
\-------------------------

Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1607; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.14)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.84)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.17)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.73)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.26)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.75)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.22)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.01)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.01)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.01)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.01)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.74)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.85)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.85)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.85)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.85)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.85)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.33)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.63)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.80)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.50)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.50)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.74)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.37)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.37)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.37)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.37)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.37)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.65)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.06)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.49)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.54)
27% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 916
\-------------------------

Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1604; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.88)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.13)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.13)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.63)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.18)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.18)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.18)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.03)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.49)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.49)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.49)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.49)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'left')
Agent drove forward instead of left. (rewarded 1.01)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.35)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.33)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.62)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.12)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.22)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.22)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.22)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.22)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.55)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 917
\-------------------------

Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1601; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.58)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.75)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.38)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.38)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.38)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.38)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.38)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.60)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 918
\-------------------------

Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1598; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.69)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.38)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.57)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.31)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.01)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.01)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.01)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.07)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.57)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.57)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.57)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.31)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.35)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.79)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.73)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.60)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', None)
Agent drove forward instead of right. (rewarded 1.22)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'left', None)
Agent drove forward instead of right. (rewarded 1.22)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.10)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.69)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 919
\-------------------------

Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1595; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 0.57)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.83)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.28)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.47)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.47)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.47)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.47)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.47)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.01)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.01)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.84)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.97)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.19)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.30)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.10)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.08)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded -0.29)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.29)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 920
\-------------------------

Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1591; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 1.78)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.60)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.60)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.60)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.60)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.60)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.20)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 2.53)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 1.65)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 1.65)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.15)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.15)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.15)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.94)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.35)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.33)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.85)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.30)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.60)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.20)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent drove right instead of forward. (rewarded 0.77)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.68)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded -0.40)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.16)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.30)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.58)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.38)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 921
\-------------------------

Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1588; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.72)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.92)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.50)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.50)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.50)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.03)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 2.33)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.46)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.32)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.52)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.26)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.26)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.67)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.97)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.69)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.63)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 922
\-------------------------

Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1585; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.01)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.56)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.35)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.35)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.35)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.35)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.35)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.35)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.35)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.06)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.94)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.93)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.49)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.47)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.47)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.10)
67% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 923
\-------------------------

Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1582; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'right')
Agent properly idled at a red light. (rewarded 2.48)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.80)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.95)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.91)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.91)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.91)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.04)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.61)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.07)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.76)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.16)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.19)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.14)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.91)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.91)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.91)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.91)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.08)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.08)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.66)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.66)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.91)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.42)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.42)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.42)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.12)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 924
\-------------------------

Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1579; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.87)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.56)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.19)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.83)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.63)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.27)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.97)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.09)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.18)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.20)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.16)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.16)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.53)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.79)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.79)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.79)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.79)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.91)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.32)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.87)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.13)
27% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 925
\-------------------------

Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1576; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 3.00)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.88)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent drove forward instead of right. (rewarded 1.91)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.24)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.80)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.11)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.59)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 926
\-------------------------

Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1572; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.19)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.09)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.09)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.09)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.09)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.09)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.84)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.72)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.72)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.67)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.67)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.67)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.67)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.42)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.36)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.48)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.48)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.48)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.48)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.48)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.48)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.48)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.48)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.48)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.48)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.48)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.64)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 927
\-------------------------

Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1569; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.60)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.18)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 1.68)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.72)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.21)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.60)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.87)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.49)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.17)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.64)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.64)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.64)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.88)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 928
\-------------------------

Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1566; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.50)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.94)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.57)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.64)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.37)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.46)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.60)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.01)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.78)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.78)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.78)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.78)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.73)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.22)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.22)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.22)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.22)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.22)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.04)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.26)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.55)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.24)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.04)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 929
\-------------------------

Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1563; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.83)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.09)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.40)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.38)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.44)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.69)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.88)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.99)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.54)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.77)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.77)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.77)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.50)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.08)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.59)
47% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 930
\-------------------------

Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1560; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.39)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 1.61)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.84)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.84)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.34)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.34)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.34)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.34)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.34)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.16)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 1.87)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 1.87)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.33)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.69)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.83)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.89)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.89)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.89)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.89)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.77)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.07)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.07)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.07)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.07)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.42)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 931
\-------------------------

Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1557; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.96)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.70)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.89)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.82)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.58)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.73)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.96)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.73)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.41)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.14)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 932
\-------------------------

Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1554; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.43)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'left')
Agent followed the waypoint left. (rewarded 2.85)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.26)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.96)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.58)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.35)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.11)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.11)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.11)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.11)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.38)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.31)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.22)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.22)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.22)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.20)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 933
\-------------------------

Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1551; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.94)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.98)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.30)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.02)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.02)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.73)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.49)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.99)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.43)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.80)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 0.57)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 0.57)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 0.57)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.95)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.95)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.83)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.83)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.24)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.24)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.24)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.24)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.71)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.61)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.89)
37% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.89)
37% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.89)
37% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.89)
37% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.89)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.87)
33% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 934
\-------------------------

Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1547; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.77)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 1.41)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.43)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.79)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.79)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.89)
76% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 935
\-------------------------

Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1544; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.91)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.93)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.70)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.26)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.26)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.26)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.26)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.71)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.64)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.84)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.84)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.84)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.84)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.96)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.34)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.79)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.29)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 1.55)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.11)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.18)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.57)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 936
\-------------------------

Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1541; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.46)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.75)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.96)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.49)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.49)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.49)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.23)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.23)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.23)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.47)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.86)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.75)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.26)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove right instead of forward. (rewarded 1.62)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.28)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.28)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.28)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.28)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.34)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.44)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.50)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.50)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.50)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.50)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.15)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 0.51)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.49)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.03)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.59)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.59)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.59)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.59)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.59)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.68)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.25)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 937
\-------------------------

Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1538; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.60)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.55)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.42)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.04)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.04)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.04)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.04)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.08)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.83)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.82)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.08)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.11)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.32)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 938
\-------------------------

Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1535; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.78)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.61)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.95)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.84)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.84)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.84)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.84)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.06)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.06)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.62)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 0.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 0.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 0.82)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.00)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.40)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.89)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.89)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.89)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.89)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.89)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.33)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.45)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.03)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.16)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.02)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.45)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.34)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 939
\-------------------------

Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1532; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.25)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.26)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.82)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.82)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.82)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.82)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.82)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.12)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.33)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.26)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.90)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.98)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.31)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.55)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.07)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 0.18)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.02)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.02)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.02)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.02)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.02)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 0.89)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove left instead of right. (rewarded -0.08)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.34)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.64)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.16)
46% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 940
\-------------------------

Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1529; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.28)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.89)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.83)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.39)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.39)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.89)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.29)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.65)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.60)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.32)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.73)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.73)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.42)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.57)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.57)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.57)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.57)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.57)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.67)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.93)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.93)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.93)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.85)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.85)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.03)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 941
\-------------------------

Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1526; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.80)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.94)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.43)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.12)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.12)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 0.19)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.27)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.77)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.77)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.77)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.11)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.11)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.24)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.85)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.54)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.08)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.26)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.27)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 942
\-------------------------

Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1523; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.50)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.01)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.46)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.70)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.94)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.20)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.48)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.83)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.70)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.01)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.50)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.50)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.50)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.50)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.50)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.55)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.49)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.34)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.17)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.61)
37% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.61)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.32)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.46)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.23)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.85)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.85)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.14)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.67)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.54)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.36)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.36)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.36)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.36)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.36)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.36)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.36)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.36)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.18)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.83)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 943
\-------------------------

Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1520; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent drove forward instead of right. (rewarded 1.49)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.75)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.16)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.81)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.19)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.10)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.84)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.19)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 944
\-------------------------

Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1517; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.30)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.84)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.84)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.84)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.90)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.90)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.90)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.97)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.04)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.02)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.41)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 2.43)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.46)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.46)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.46)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.64)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.59)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.77)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
49% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
49% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
49% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.51)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.40)
37% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 945
\-------------------------

Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1514; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.31)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.86)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.39)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.47)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.85)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.85)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.85)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.85)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.03)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.56)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.66)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.66)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.66)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.66)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.24)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.97)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.97)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.97)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.97)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.97)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.97)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.97)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.84)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.41)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.92)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.08)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.04)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.50)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.50)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.50)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.50)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.50)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.50)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.50)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.50)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.50)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.50)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.50)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.97)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.34)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 946
\-------------------------

Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1511; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.24)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.78)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.49)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.85)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.85)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.85)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.85)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.85)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.85)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.85)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.85)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.21)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.21)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.21)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.21)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.21)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 0.92)
72% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 947
\-------------------------

Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1508; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.45)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 0.96)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.87)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.93)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.07)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.36)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.87)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.89)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.59)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.65)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.28)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.52)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.19)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 948
\-------------------------

Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1505; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 2.23)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.61)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.54)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.77)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.69)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.92)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.88)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.90)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.90)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.90)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.90)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.65)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.79)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 949
\-------------------------

Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1502; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.02)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.65)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.80)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.51)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.39)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.39)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.39)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.39)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.48)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.91)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.78)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.45)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.64)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.64)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.64)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.64)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.64)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.38)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.93)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.91)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.91)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.91)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.91)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.92)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 950
\-------------------------

Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1499; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded 0.71)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent drove right instead of left. (rewarded 0.32)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.43)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.15)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.21)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.21)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.21)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.21)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.21)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.51)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.74)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.14)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.62)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.72)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.01)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.01)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.51)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 951
\-------------------------

Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1496; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.42)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.35)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.39)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.39)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.39)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.39)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.39)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.39)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.53)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.87)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.83)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.73)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.58)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.95)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.95)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.44)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.44)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.44)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.64)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.36)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.73)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.89)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.08)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.08)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.08)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.08)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.08)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.86)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.49)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.61)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.80)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 952
\-------------------------

Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1493; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.20)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.02)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.15)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.28)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.68)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.21)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.28)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 953
\-------------------------

Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1490; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.99)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.97)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.42)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.05)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.78)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.85)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.90)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.10)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.10)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.37)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.81)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.21)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent followed the waypoint forward. (rewarded 2.06)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 954
\-------------------------

Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1487; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.18)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 0.98)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.58)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.54)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.74)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.53)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.14)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 1.21)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.26)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.56)
67% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 955
\-------------------------

Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1484; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.24)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.21)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.56)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.56)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.56)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.80)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.54)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.17)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.28)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.28)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.28)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.66)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.66)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.86)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 956
\-------------------------

Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1481; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.05)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.35)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.71)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.15)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.53)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 2.52)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 2.52)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 2.52)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.35)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.51)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.51)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.27)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 957
\-------------------------

Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1478; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.58)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent drove right instead of forward. (rewarded 0.50)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 1.73)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.05)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.04)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.04)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.65)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.84)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.92)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.76)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.11)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.56)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.08)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.30)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.30)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.24)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.24)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.97)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.11)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.11)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 958
\-------------------------

Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1475; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.34)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.06)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.90)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.38)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.69)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.32)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.10)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.56)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.76)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.76)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.76)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.76)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.84)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.76)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.79)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.73)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.37)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.37)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.37)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.37)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.07)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.25)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.69)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'right')
Agent followed the waypoint forward. (rewarded 2.21)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'right')
Agent followed the waypoint forward. (rewarded 2.21)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.45)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.31)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.16)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 959
\-------------------------

Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1472; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.54)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.73)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.30)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'right')
Agent drove forward instead of right. (rewarded 0.61)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.08)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.08)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.57)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.55)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.96)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.52)
57% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 960
\-------------------------

Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1469; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', None)
Agent drove forward instead of right. (rewarded 1.71)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.85)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.69)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.69)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.13)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.17)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.14)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.39)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.18)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.18)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 0.88)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.60)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.60)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.72)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.99)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.91)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.91)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.94)
27% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.94)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.44)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.44)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.44)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.69)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
17% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
17% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
17% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
17% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.30)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.12)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.76)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.27)
3% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 961
\-------------------------

Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1466; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.05)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.20)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.39)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.54)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.83)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.68)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.98)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 962
\-------------------------

Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1463; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent drove forward instead of left. (rewarded 1.02)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.91)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.72)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.39)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.64)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.31)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded -0.08)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded -0.08)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded -0.08)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded -0.08)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded -0.08)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded -0.08)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.41)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.72)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.72)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.72)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.72)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.27)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.64)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.12)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.12)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.12)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.12)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.67)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'right')
Agent drove forward instead of left. (rewarded 0.98)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', 'right')
Agent drove forward instead of left. (rewarded 0.98)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', 'right')
Agent drove forward instead of left. (rewarded 0.98)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.46)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.46)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.21)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 963
\-------------------------

Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1460; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.73)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.93)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.17)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.31)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.84)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.04)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.44)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.91)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.97)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.73)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.73)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.73)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.73)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.73)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.73)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.73)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.50)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 964
\-------------------------

Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1457; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.00)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.36)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.65)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.50)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.55)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.62)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.64)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.73)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 965
\-------------------------

Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1454; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.64)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.69)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent drove right instead of forward. (rewarded 1.04)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.87)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.19)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.03)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.04)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.04)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.04)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.04)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.59)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.34)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.15)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.15)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.67)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded -0.16)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.17)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.17)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.17)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.17)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.35)
47% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 966
\-------------------------

Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1451; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.41)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.48)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.80)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.84)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.97)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.96)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.13)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.13)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.93)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.13)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.79)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.22)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.20)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.20)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.20)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.72)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 967
\-------------------------

Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1449; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.68)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.91)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.19)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.51)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.96)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.04)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.08)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.99)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.62)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.41)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.63)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.63)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.34)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.34)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.34)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.74)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.74)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.26)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.98)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.98)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.98)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.98)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.98)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.98)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.51)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.92)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.90)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.01)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.78)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.80)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.12)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.32)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.09)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 968
\-------------------------

Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1446; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.65)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.13)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.65)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.65)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.65)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.65)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.94)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.89)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 2.63)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 2.63)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.16)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.64)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.21)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.79)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.79)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.79)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.79)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.66)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.19)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.01)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.01)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.01)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.01)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.01)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.08)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.08)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.08)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.41)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.25)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 969
\-------------------------

Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1443; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.69)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.79)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.97)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.74)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.96)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.85)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.35)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.16)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.16)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.16)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.16)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.86)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.32)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 970
\-------------------------

Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1440; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.61)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.07)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.65)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.27)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.66)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.01)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.69)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.00)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.00)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.00)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.00)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.99)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.40)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.40)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.48)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.06)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.31)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.23)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.59)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.85)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.59)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.20)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.15)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.52)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 971
\-------------------------

Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1437; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.95)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.86)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.85)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.60)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.31)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.34)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.34)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.50)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.40)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.40)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.46)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.51)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
63% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 972
\-------------------------

Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1434; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.71)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.49)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.60)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.58)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.36)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.76)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.15)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 973
\-------------------------

Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1431; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.63)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.23)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.60)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.60)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.79)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.05)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.05)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.39)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.22)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove left instead of right. (rewarded 1.37)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.56)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.39)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.39)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.39)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.39)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.00)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.07)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.43)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.43)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.07)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.07)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.07)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.83)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.65)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 974
\-------------------------

Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1428; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.94)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.46)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.18)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.74)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.19)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.63)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.63)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.63)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.63)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.65)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.30)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.60)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.60)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.36)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.79)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.52)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.30)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.62)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.62)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.80)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 975
\-------------------------

Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1426; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.71)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.27)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.72)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.72)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.72)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.72)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.64)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.64)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.69)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 976
\-------------------------

Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1423; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.47)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.71)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.00)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.00)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.16)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.84)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.15)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.15)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.15)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.15)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.15)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.15)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.34)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.56)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.72)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.70)
63% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 977
\-------------------------

Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1420; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.67)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.21)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.33)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.37)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.52)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.87)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 978
\-------------------------

Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1417; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.76)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.67)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.09)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.88)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.73)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.81)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.97)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.97)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.97)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.97)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.97)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.37)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.13)
48% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.13)
48% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.13)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.96)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 979
\-------------------------

Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1414; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.24)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.37)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.58)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.33)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.15)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 980
\-------------------------

Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1411; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'right')
Agent drove left instead of right. (rewarded 0.89)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.10)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.10)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.10)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.10)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.10)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.10)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.10)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.03)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.00)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.04)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.73)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.06)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.06)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.53)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.53)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.53)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.96)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.11)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.66)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.52)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.67)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.21)
43% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 981
\-------------------------

Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1409; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.47)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.15)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.15)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.15)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.15)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.15)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.13)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.51)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.31)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.56)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.42)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.53)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.38)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.20)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded -0.06)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.29)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.29)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.29)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.29)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.01)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.36)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.36)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.36)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.36)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.36)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.36)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.36)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.59)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.59)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 2.42)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.06)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.25)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.06)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.05)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.99)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.66)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.09)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.03)
3% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 982
\-------------------------

Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1406; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent properly idled at a red light. (rewarded 2.17)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.35)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.01)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.01)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.93)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.93)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.83)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.42)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.72)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.67)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 1.60)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.27)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.44)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 983
\-------------------------

Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1403; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.85)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.83)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.51)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.51)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.51)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.51)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.51)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.51)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 1.91)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.32)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.72)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.04)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.15)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.92)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.57)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.57)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.57)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.73)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.95)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.16)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.43)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.83)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.83)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.83)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.79)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 0.84)
24% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 984
\-------------------------

Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1400; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.31)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.88)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.87)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.94)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 1.91)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.50)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.50)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.97)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.82)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.23)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.07)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 985
\-------------------------

Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1397; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.92)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.24)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.42)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.14)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.14)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.14)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.14)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.79)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.44)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.69)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 986
\-------------------------

Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1395; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.61)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.47)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.84)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.23)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.95)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 987
\-------------------------

Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1392; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.05)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.65)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.74)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.74)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.74)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.74)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.34)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.34)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.61)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.33)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.23)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.84)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.98)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.70)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.21)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.45)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.25)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.02)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.57)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.48)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 988
\-------------------------

Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1389; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.91)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.97)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.17)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.17)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.17)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.17)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.17)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.75)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.22)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.00)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.00)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.00)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.75)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.00)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.52)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.18)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.78)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.78)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded -0.16)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.15)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.33)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 989
\-------------------------

Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1386; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.93)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.30)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.82)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.82)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.82)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.82)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.82)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.82)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.82)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.82)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.82)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.82)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.82)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.59)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.12)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.52)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.52)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.52)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.40)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.51)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.97)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.94)
63% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 990
\-------------------------

Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1383; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent drove forward instead of left. (rewarded 0.73)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.83)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.85)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.96)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 1.77)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.79)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.36)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.24)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.01)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.72)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.21)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.28)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.88)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 991
\-------------------------

Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1381; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.72)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.83)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.91)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.91)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.91)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.91)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.32)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.32)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.76)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.76)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.44)
76% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 992
\-------------------------

Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1378; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.47)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.37)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.58)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.52)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.92)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.66)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.81)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.81)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.81)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.97)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.76)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.62)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.50)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.50)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.96)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 993
\-------------------------

Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1375; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.38)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.04)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.54)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.11)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.78)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.71)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.81)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.81)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.81)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.34)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.09)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 994
\-------------------------

Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1372; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.88)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.04)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.08)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.08)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.08)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.08)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.08)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.25)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.18)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.71)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.11)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.43)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.68)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.90)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.90)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.90)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.91)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.76)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 995
\-------------------------

Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1370; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.22)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.98)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.46)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.46)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.46)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.46)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.46)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.02)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.21)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.40)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 0.20)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 0.20)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.74)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.74)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.28)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.04)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.79)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.79)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.79)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.79)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 1.91)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.72)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.30)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.30)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.30)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.30)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.30)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.42)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.42)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.42)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.42)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.09)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.09)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.09)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.85)
27% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 996
\-------------------------

Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1367; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 1.85)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.14)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove left instead of forward. (rewarded 0.81)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove left instead of forward. (rewarded 0.81)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove left instead of forward. (rewarded 0.81)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove left instead of forward. (rewarded 0.81)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.90)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.35)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.27)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.35)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.31)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 997
\-------------------------

Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1364; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.57)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.52)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.31)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.13)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.16)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.56)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.02)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.63)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.53)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 998
\-------------------------

Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1361; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.33)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.88)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.50)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.52)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.38)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.54)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.43)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.53)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'right')
Agent followed the waypoint forward. (rewarded 2.45)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 999
\-------------------------

Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1359; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.01)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.71)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.47)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.47)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.47)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.47)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.90)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.72)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.08)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.65)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.01)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.01)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.01)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.79)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.79)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.79)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.79)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.34)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.30)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.36)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 0.75)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.80)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.80)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.36)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1000
\-------------------------

Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1356; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.32)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.72)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.56)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.31)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.65)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.72)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.72)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.72)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.72)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.43)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.59)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.73)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.07)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.46)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.20)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.16)
57% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1001
\-------------------------

Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1353; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.47)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.80)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.97)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.42)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.10)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.23)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.66)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.66)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.66)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.47)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.47)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.47)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.23)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.60)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.46)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1002
\-------------------------

Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1351; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.19)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove left instead of forward. (rewarded 0.19)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.47)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.59)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.04)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.32)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.32)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.61)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1003
\-------------------------

Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1348; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.28)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded -0.00)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.36)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.60)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.27)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.27)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.27)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.27)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.98)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.82)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.82)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.82)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.82)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.84)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded -0.06)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded -0.06)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded -0.06)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded -0.06)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.58)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1004
\-------------------------

Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1345; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.67)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.61)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.13)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.33)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.38)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.70)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.70)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.70)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.70)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.64)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.31)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.63)
63% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1005
\-------------------------

Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1343; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.65)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.47)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.76)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.32)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.43)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.46)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.28)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.06)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.35)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.87)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.62)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.54)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.83)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1006
\-------------------------

Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1340; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.12)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.54)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.01)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.53)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.23)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.26)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.47)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.52)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.01)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.01)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.01)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.67)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1007
\-------------------------

Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1337; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.22)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.62)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove left instead of forward. (rewarded 1.35)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.42)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.31)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.90)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.74)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.06)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.44)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.88)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.84)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1008
\-------------------------

Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1335; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.65)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.88)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.87)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.01)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.17)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.90)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1009
\-------------------------

Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1332; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.45)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.36)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.65)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.65)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.83)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.32)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.32)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.32)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.32)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.32)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.32)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.98)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.50)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.23)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.16)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.16)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.67)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.67)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.67)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.56)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.57)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.77)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.50)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.87)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.67)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.04)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.35)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.35)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.35)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.49)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.14)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.20)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.57)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.47)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded -0.36)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded -0.36)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded -0.36)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.98)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1010
\-------------------------

Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1329; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.16)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.41)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove forward instead of right. (rewarded 1.94)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove forward instead of right. (rewarded 1.94)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove forward instead of right. (rewarded 1.94)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.21)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.43)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.43)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.43)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.43)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.23)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.23)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.04)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.25)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.25)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.07)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.07)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.07)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.77)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.77)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.77)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.77)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.09)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1011
\-------------------------

Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1327; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.84)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.81)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.83)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.35)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.85)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.85)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.85)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.85)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.31)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.11)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.86)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.75)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.75)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.75)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.03)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.03)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.97)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.34)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.59)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'right')
Agent drove right instead of forward. (rewarded 0.23)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.77)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1012
\-------------------------

Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1324; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.63)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.84)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.52)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.78)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.90)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.37)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.64)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.54)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.70)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.70)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.70)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.70)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.70)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.70)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.70)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.70)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.63)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.06)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.12)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1013
\-------------------------

Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1321; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.26)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.61)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.52)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.77)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.15)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.80)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1014
\-------------------------

Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1319; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.40)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.76)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.22)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.25)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.12)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.12)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.05)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.67)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.27)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.35)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.35)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.35)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.35)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.97)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.60)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.52)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.40)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.63)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.11)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.30)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.10)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.16)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.16)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.23)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.23)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.23)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.23)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.23)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.23)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.23)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded -0.52)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.36)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.36)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.36)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.36)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.36)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.36)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.36)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.36)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.36)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.36)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.36)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded 1.09)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.64)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.64)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.64)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.64)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.64)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.64)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.64)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.64)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.64)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.64)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.64)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.42)
4% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1015
\-------------------------

Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1316; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.46)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.86)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.53)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.51)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.51)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.51)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.51)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.51)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.51)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.51)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.51)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.51)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.51)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.51)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.95)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.41)
73% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1016
\-------------------------

Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1313; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.94)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.52)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.52)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.52)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.50)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.46)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.37)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.79)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.74)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.06)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.06)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.06)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.06)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.25)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.31)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.31)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.49)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.98)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.98)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.31)
36% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1017
\-------------------------

Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1311; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.08)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.62)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.44)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.44)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.44)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.44)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 1.26)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.62)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.87)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.22)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.39)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.39)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.73)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.13)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.87)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.87)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.87)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.87)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.87)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.26)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.74)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1018
\-------------------------

Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1308; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.96)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.23)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.60)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.60)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.60)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.60)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.60)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.60)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.60)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.60)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.10)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.97)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.26)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.45)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.35)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.23)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.23)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.43)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.69)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.91)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.97)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.09)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.65)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1019
\-------------------------

Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1305; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.00)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.80)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.55)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.84)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.88)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.88)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.69)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.76)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.76)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.76)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.28)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 0.91)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 0.96)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.97)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.97)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.82)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1020
\-------------------------

Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1303; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.72)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.23)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.65)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.90)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.20)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.20)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.20)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.20)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.20)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.17)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.19)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.40)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.70)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.70)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.70)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.70)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.70)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.70)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.70)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.69)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.44)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.44)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.44)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.17)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.66)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.71)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.71)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.71)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.71)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.77)
5% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1021
\-------------------------

Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1300; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.88)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.46)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.14)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.16)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 1.44)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.54)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.50)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.09)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1022
\-------------------------

Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1298; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.26)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.63)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.14)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.31)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.05)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.43)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.28)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.20)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.20)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.08)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.66)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1023
\-------------------------

Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1295; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.19)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.51)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.12)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.83)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1024
\-------------------------

Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1293; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.18)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.22)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.40)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.49)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1025
\-------------------------

Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1290; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.21)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.51)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.81)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.39)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.12)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.81)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.60)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.17)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.59)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.37)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.13)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.72)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.72)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.72)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.72)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.12)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.66)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.66)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.42)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.59)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.59)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.59)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.59)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.59)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded -0.17)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.11)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.10)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.10)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.65)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.28)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.28)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.28)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.17)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.99)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1026
\-------------------------

Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1287; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.08)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.62)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.72)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.22)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.84)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.28)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.42)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.42)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.82)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.53)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.61)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.39)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.44)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.47)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.79)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.79)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.79)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.59)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.59)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.05)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.36)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.83)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.54)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.58)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.65)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.65)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.65)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.65)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.84)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.22)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.99)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.34)
3% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1027
\-------------------------

Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1285; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.38)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.35)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.04)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.26)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.26)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.86)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.03)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.89)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.35)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.64)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent followed the waypoint left. (rewarded 1.20)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.30)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.39)
66% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1028
\-------------------------

Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1282; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.24)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.68)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.18)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.72)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.53)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.12)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.09)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.90)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.34)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.58)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.58)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.61)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.61)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.61)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.11)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.11)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.11)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.11)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.11)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.20)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1029
\-------------------------

Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1280; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.24)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 1.10)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.36)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.05)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.12)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.83)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.18)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.30)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.41)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.24)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.30)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1030
\-------------------------

Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1277; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.80)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.68)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.11)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.77)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.04)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.30)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 2.91)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.81)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.32)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.21)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.19)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1031
\-------------------------

Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1275; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.49)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.26)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.83)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.06)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.15)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.58)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.04)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.90)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.88)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1032
\-------------------------

Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1272; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.88)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 1.17)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.91)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.05)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.46)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.46)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.46)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.46)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.89)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 2.06)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 2.06)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 2.06)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.83)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.88)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.92)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.22)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.26)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1033
\-------------------------

Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1269; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.37)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.76)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.76)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.76)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.76)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.13)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.42)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.42)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.42)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.33)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1034
\-------------------------

Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1267; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.43)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.23)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.23)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.23)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.23)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.23)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.23)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.15)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.45)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.45)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.45)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.03)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.35)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.62)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.62)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.62)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.62)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent followed the waypoint left. (rewarded 2.29)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.76)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.17)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1035
\-------------------------

Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1264; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.88)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.39)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.63)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.87)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.87)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.87)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.04)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1036
\-------------------------

Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1262; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.29)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.90)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.00)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.00)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.00)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.00)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.11)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.31)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.71)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.74)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.74)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.29)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.40)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.40)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.40)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.96)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.59)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.18)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.89)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.93)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.37)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.37)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.37)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.37)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.96)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.61)
37% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1037
\-------------------------

Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1259; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.85)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.75)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.64)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.64)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.37)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.37)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.13)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.71)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.42)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.96)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.66)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.47)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.89)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.43)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1038
\-------------------------

Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1257; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.77)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.21)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.71)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.71)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.71)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.71)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.71)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.71)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.71)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.92)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.64)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.60)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.21)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1039
\-------------------------

Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1254; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 1.32)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.27)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.05)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.05)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.30)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.54)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 2.78)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 2.78)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 2.78)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 2.78)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.13)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.24)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.24)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.24)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.12)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1040
\-------------------------

Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1252; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.72)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.99)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.51)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.56)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.90)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.90)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.90)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.78)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1041
\-------------------------

Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1249; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.90)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.43)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.12)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.50)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.10)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.07)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.91)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.23)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.23)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.23)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.23)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.37)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.42)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.42)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.42)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.42)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.73)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.81)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.79)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.01)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1042
\-------------------------

Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1247; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.97)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.80)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.73)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.73)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.73)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.93)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.25)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.81)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.06)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.32)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1043
\-------------------------

Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1244; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.37)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.47)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.28)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1044
\-------------------------

Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1242; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.38)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.94)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.52)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.52)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.52)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.52)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.38)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.35)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.35)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.16)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.16)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.16)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.22)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.93)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.93)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.93)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.93)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.22)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.94)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.13)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.58)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.58)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.58)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.58)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.39)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1045
\-------------------------

Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1239; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.48)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.71)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.71)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.71)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.71)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.71)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.39)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.85)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.15)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.23)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.03)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.00)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1046
\-------------------------

Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1237; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.69)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.58)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.40)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.40)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.40)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.40)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.40)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.87)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.51)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.51)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.51)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.51)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.32)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.03)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.54)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.17)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.90)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.61)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.61)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.61)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.61)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.14)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1047
\-------------------------

Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1234; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.57)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.54)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.14)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.81)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.81)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.32)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.31)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded -0.08)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.09)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.02)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.99)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.99)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.99)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.99)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.25)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.78)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.47)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.47)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.47)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.34)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.34)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.51)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.14)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.99)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.01)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.01)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.01)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.41)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.06)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1048
\-------------------------

Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1232; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.77)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent drove right instead of left. (rewarded 0.22)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.50)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.50)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.50)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.47)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.09)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.09)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.22)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.26)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.89)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.89)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.89)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove forward instead of right. (rewarded 1.61)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.59)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.33)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.21)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1049
\-------------------------

Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1229; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.14)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.61)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.14)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.14)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.14)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.29)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.26)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.28)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.57)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.74)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.23)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.32)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.62)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.62)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.62)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.62)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.62)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.80)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.05)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.05)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.88)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.96)
5% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.96)
5% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.96)
5% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.96)
5% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.96)
5% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.96)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.37)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1050
\-------------------------

Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1227; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.64)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.80)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.70)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.09)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.80)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.80)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.80)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.80)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.64)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.64)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.64)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.64)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.50)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.88)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.88)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.88)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.40)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.70)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.04)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.77)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.83)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.83)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.88)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.11)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.11)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1051
\-------------------------

Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1225; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.79)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.72)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.28)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.82)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.53)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.53)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.53)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.69)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.29)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.19)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.90)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.26)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.26)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.26)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.26)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.82)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.62)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.62)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.62)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.62)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.62)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.45)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1052
\-------------------------

Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1222; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.31)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.19)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.34)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.94)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.08)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.56)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.69)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1053
\-------------------------

Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1220; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.98)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.87)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.87)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.23)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.96)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.96)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.96)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.47)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.30)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.96)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.96)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.41)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.58)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.41)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.00)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.31)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.56)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.59)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.59)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.59)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.59)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.66)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.17)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1054
\-------------------------

Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1217; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.45)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.35)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.85)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.70)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.70)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.60)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.60)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.60)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.71)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.36)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.07)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.84)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.10)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.35)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.76)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.95)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1055
\-------------------------

Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1215; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.58)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.86)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.58)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.73)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.44)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.82)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.24)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.42)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1056
\-------------------------

Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1212; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.84)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.13)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.81)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.26)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.74)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.16)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.11)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.91)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.47)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.88)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.82)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.73)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.81)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.00)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.82)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.76)
54% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1057
\-------------------------

Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1210; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.25)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.93)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.75)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.97)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.97)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.97)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.97)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.97)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.97)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.97)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.97)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.97)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.97)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.97)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.97)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.97)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.26)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.86)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.05)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 1.19)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.35)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.50)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.38)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.38)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.38)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.38)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.67)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 0.74)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.55)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.04)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.04)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.04)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.04)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.04)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.99)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.60)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.60)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.60)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.60)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.60)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.60)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.60)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.15)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1058
\-------------------------

Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1208; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.65)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.72)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.47)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.71)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.76)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.47)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.18)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.97)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.92)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.91)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.91)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.91)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.91)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.05)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.21)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.21)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent followed the waypoint right. (rewarded 1.18)
48% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent followed the waypoint right. (rewarded 1.18)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.82)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.82)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.82)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.27)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.46)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.46)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.46)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.46)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.46)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.04)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.58)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.46)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.06)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.14)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.54)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1059
\-------------------------

Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1205; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.15)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.95)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.80)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.80)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.80)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.80)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.53)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.19)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.62)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.10)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.33)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.73)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.75)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent drove right instead of forward. (rewarded 1.42)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent drove right instead of forward. (rewarded 1.42)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent drove right instead of forward. (rewarded 1.42)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent drove right instead of forward. (rewarded 1.42)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.57)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.07)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.87)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.56)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.49)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 0.90)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.16)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1060
\-------------------------

Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1203; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.10)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.49)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.83)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 2.28)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.53)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.53)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.36)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1061
\-------------------------

Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1200; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.87)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.18)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.43)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.33)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove right instead of forward. (rewarded 1.11)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove right instead of forward. (rewarded 1.11)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove right instead of forward. (rewarded 1.11)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove right instead of forward. (rewarded 1.11)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove right instead of forward. (rewarded 1.11)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.57)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.13)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.13)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.13)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1062
\-------------------------

Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1198; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.76)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.00)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.30)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.14)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.20)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.18)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1063
\-------------------------

Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1196; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 0.31)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.49)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.69)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.69)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.69)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.69)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.69)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.43)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.18)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.18)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.18)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.18)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.18)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.18)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.20)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.49)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.88)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.37)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.81)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.81)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.81)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.27)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.96)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.20)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.23)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.84)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.14)
37% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1064
\-------------------------

Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1193; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.84)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.25)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.27)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.89)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.95)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.95)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.95)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.95)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.00)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.00)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.17)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.17)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.20)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.43)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.73)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.40)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 0.41)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1065
\-------------------------

Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1191; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.25)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.61)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.24)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.20)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.20)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.29)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.78)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.20)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.49)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.50)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.50)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.50)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.50)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.04)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.80)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.44)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.35)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 0.92)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.71)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.36)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1066
\-------------------------

Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1188; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.04)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.85)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.70)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.56)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 1.06)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.39)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.39)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.88)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.88)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.88)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.88)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.62)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.62)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.62)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.62)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.23)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.06)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.59)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1067
\-------------------------

Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1186; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.54)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 2.98)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.64)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.12)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.90)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.93)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.93)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.93)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.60)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.12)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.12)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.12)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.12)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.11)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded 0.77)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.15)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.15)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.15)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.48)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.74)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.08)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 0.86)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.17)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.83)
28% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.83)
28% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.83)
28% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.83)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.02)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove forward instead of right. (rewarded 1.00)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.77)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.31)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.74)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.74)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.74)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.74)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.74)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 0.14)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1068
\-------------------------

Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1184; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'right')
Agent followed the waypoint forward. (rewarded 2.59)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.21)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.21)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.21)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 1.57)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.19)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.19)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.85)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1069
\-------------------------

Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1181; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.28)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.31)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.52)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.12)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.53)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.40)
72% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1070
\-------------------------

Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1179; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.25)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.26)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.65)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent drove right instead of left. (rewarded 0.29)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent drove right instead of left. (rewarded 0.29)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.06)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.40)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.64)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1071
\-------------------------

Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1177; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove left instead of right. (rewarded 0.85)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.93)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.23)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.23)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.23)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.23)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.23)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.53)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.12)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.57)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.37)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.62)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.26)
63% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1072
\-------------------------

Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1174; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.15)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.71)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.98)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.88)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.77)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.77)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.53)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.89)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.12)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.12)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.12)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.12)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.12)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.21)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.76)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1073
\-------------------------

Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1172; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.87)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.61)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.43)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.11)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.38)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.59)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.76)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.19)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1074
\-------------------------

Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1170; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.27)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'right')
Agent properly idled at a red light. (rewarded 1.30)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.60)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.31)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.05)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.05)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.21)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.20)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.14)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 1.29)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded -0.05)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.33)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.72)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.28)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.28)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.28)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.16)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.24)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.24)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.18)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.18)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.60)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.60)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.60)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.79)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.53)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.23)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.71)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.71)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.62)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.62)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.05)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.69)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.95)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded -0.64)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1075
\-------------------------

Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1167; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.10)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.71)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'right')
Agent followed the waypoint forward. (rewarded 1.66)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.27)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.68)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.92)
77% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1076
\-------------------------

Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1165; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.93)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.57)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.25)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.91)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.25)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.20)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.29)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.73)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.37)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.52)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.13)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.43)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.43)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.43)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.43)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded -0.23)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded -0.23)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.80)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.92)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.13)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.86)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.86)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.86)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.86)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.51)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.23)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.36)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.52)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.18)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.18)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.18)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.18)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1077
\-------------------------

Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1163; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.35)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.47)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.43)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.08)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.08)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.08)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.08)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.08)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.34)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.60)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.60)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.60)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.60)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.60)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.64)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.48)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.48)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.48)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.55)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.35)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.56)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.21)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.21)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.21)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.79)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.90)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.90)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.90)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.90)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.06)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.15)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.31)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.15)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.15)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.15)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.47)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1078
\-------------------------

Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1160; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.21)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.23)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.47)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.91)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.58)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.54)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.54)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.54)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.82)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.54)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.15)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.15)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.15)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.22)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.33)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1079
\-------------------------

Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1158; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.37)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.71)
93% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.71)
93% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.71)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.63)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.86)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.02)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.78)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.65)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 0.63)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.50)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.97)
47% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1080
\-------------------------

Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1156; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.02)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.85)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 2.06)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.47)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.53)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.71)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.42)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.97)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1081
\-------------------------

Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1153; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.89)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.41)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.71)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.71)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.72)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.72)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.72)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.72)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.72)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.72)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.72)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.80)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.85)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.54)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.21)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1082
\-------------------------

Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1151; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.18)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.61)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.61)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.61)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.61)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.61)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.61)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.60)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.14)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.14)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.14)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
71% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
71% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.08)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.09)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.22)
63% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1083
\-------------------------

Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1149; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.99)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.20)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.22)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.14)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.29)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.29)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.66)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.35)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.02)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.54)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.54)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.54)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.54)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.54)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.02)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1084
\-------------------------

Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1146; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.19)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.31)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.71)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.72)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.72)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.00)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.00)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.00)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.38)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.44)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.44)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.44)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.44)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.41)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.61)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.12)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.63)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent drove forward instead of left. (rewarded 1.08)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.90)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.26)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.26)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.26)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.26)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 0.43)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.44)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.93)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.34)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.34)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.34)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.34)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.12)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.94)
13% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1085
\-------------------------

Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1144; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.52)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.81)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.81)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.81)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.81)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.88)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.98)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.98)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.98)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.98)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.98)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.41)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.72)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.72)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.62)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.14)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.40)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.24)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.24)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.24)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.24)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.24)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.16)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.75)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1086
\-------------------------

Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1142; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.17)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.82)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.05)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.65)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.65)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.88)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.88)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'right')
Agent drove right instead of forward. (rewarded 0.62)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', 'right')
Agent drove right instead of forward. (rewarded 0.62)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.23)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.52)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.89)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.68)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.61)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.61)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.73)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.69)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.69)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.78)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.78)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.78)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 0.65)
5% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1087
\-------------------------

Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1139; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.64)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.24)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.23)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.41)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.80)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.93)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.71)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.87)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.50)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.51)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1088
\-------------------------

Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1137; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.94)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.50)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.50)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.50)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.50)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.50)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.50)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.54)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.58)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.18)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.94)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.60)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.24)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1089
\-------------------------

Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1135; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.42)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.65)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.85)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.29)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.29)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.29)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.51)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1090
\-------------------------

Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1133; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.78)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.28)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.28)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.94)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.94)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.94)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.94)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.09)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.79)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.59)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.59)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.59)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.59)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.61)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.49)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.68)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.68)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.68)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.68)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.68)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.30)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.21)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.21)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.21)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.21)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.27)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.79)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.57)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1091
\-------------------------

Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1130; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.29)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.68)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.52)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.95)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.52)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1092
\-------------------------

Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1128; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.20)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.93)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.76)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.06)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.71)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.71)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.95)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.14)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1093
\-------------------------

Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1126; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.56)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.83)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.95)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.99)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.85)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.85)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.85)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.85)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.38)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.11)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.11)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.11)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.11)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.11)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.11)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.11)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.73)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.29)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1094
\-------------------------

Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1124; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.21)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.99)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.99)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.99)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.99)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.99)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.99)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.33)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.81)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.15)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 0.92)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.82)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.04)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.04)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.04)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.04)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.36)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1095
\-------------------------

Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1121; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.70)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.28)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.22)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.22)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.22)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.22)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.22)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.82)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.49)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.00)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.41)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.41)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.41)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.16)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.75)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.75)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.75)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.45)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.45)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.45)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.08)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.64)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.90)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.68)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.68)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.68)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.68)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.26)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded -0.55)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded -0.55)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.49)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.76)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1096
\-------------------------

Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1119; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.56)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.14)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.49)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.76)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.08)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.45)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.66)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.15)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.31)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.35)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1097
\-------------------------

Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1117; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.58)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.54)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.71)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.04)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.85)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.85)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.85)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.15)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove left instead of forward. (rewarded 1.06)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.35)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.35)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.35)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.35)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.35)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.35)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.55)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.98)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.64)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1098
\-------------------------

Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1115; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.25)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.43)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.71)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.00)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.96)
72% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1099
\-------------------------

Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1112; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.03)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.49)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.47)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.45)
84% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.45)
84% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.45)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', None)
Agent drove forward instead of right. (rewarded 0.76)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.75)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.86)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.13)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.95)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.92)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.92)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.92)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.88)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.63)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.43)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1100
\-------------------------

Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1110; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.54)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.62)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.75)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.06)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.63)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.63)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.63)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.63)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.63)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.63)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.63)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.63)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.63)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.39)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.59)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.59)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.99)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.46)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.97)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.29)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.64)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 0.85)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.50)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1101
\-------------------------

Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1108; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.95)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.12)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.85)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.08)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.02)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.40)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.68)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.59)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1102
\-------------------------

Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1106; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.28)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.69)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.88)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.81)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.71)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.71)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.71)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.71)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.71)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.10)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.11)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.01)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.76)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.41)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.23)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1103
\-------------------------

Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1104; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.73)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.34)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.34)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.34)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.34)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.87)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.87)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.52)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.43)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.43)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.43)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.43)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.50)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.50)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.50)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.50)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.86)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1104
\-------------------------

Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1101; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.93)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.59)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.45)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.75)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.83)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.31)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.31)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.31)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 1.21)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 1.21)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 1.21)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 1.21)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.40)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove right instead of left. (rewarded 1.12)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove right instead of left. (rewarded 1.12)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove right instead of left. (rewarded 1.12)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove right instead of left. (rewarded 1.12)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.56)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.36)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.96)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'right')
Agent drove right instead of forward. (rewarded 0.25)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', 'right')
Agent drove right instead of forward. (rewarded 0.25)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.11)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 0.22)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1105
\-------------------------

Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1099; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.12)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.93)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.42)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.42)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.42)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.42)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.42)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.42)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.89)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.55)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.55)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.55)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.82)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.94)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.23)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1106
\-------------------------

Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1097; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.76)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.35)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.20)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.20)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.20)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.20)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.20)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.29)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.94)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.94)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.94)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.07)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.07)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.07)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1107
\-------------------------

Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1095; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.32)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.28)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.17)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.25)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.12)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.12)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.81)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1108
\-------------------------

Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1093; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.94)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.01)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.01)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.01)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.01)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.01)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.13)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.80)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.83)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.54)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.64)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.11)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.39)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.39)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.90)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 0.96)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1109
\-------------------------

Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1090; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.03)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.61)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.69)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.91)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.84)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.84)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.84)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.84)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.84)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.84)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.87)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.63)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.63)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.63)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.63)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.63)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.13)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.99)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.31)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.67)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.67)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.67)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.67)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.07)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.57)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.57)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded -0.14)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.91)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 0.70)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.01)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.01)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.53)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.53)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.53)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.53)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.83)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.14)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 1.74)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.38)
17% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1110
\-------------------------

Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1088; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.52)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.49)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.52)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.33)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.33)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.33)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.13)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.13)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.11)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.53)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.79)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1111
\-------------------------

Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1086; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.96)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.11)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.70)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.63)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.63)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.63)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.63)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.63)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.64)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.64)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.79)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent followed the waypoint forward. (rewarded 2.42)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.90)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1112
\-------------------------

Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1084; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.95)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.90)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.44)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.11)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.98)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.36)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.74)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.49)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.91)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.06)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.06)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.63)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.40)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.40)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.40)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded -0.05)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.58)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.03)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.52)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.58)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.58)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.58)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.58)
5% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1113
\-------------------------

Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1082; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.66)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.31)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.60)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.90)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.08)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.83)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.83)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.83)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.83)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.54)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.60)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.32)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.25)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.25)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.35)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.35)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.35)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.35)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.64)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.16)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.76)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 0.43)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.33)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.33)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.33)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.68)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.77)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.85)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.28)
16% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1114
\-------------------------

Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1080; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.24)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.79)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.58)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.94)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.94)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.20)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.10)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.88)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.79)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.79)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.79)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.52)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.22)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.27)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.60)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1115
\-------------------------

Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1077; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.72)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.14)
90% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.14)
90% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.14)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.68)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.68)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.52)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.40)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.86)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.42)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.73)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.73)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.08)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1116
\-------------------------

Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1075; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.32)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.11)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.63)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.63)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.42)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.50)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.53)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.76)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.06)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.06)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.06)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.06)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.45)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.73)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.73)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.62)
48% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.62)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.77)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.77)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.20)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.20)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.20)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.20)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.20)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.20)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.05)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.61)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.61)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'right')
Agent followed the waypoint right. (rewarded 2.14)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'right')
Agent followed the waypoint right. (rewarded 2.14)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'right')
Agent followed the waypoint right. (rewarded 2.14)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'right')
Agent followed the waypoint right. (rewarded 2.14)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'right')
Agent followed the waypoint right. (rewarded 2.14)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.69)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.90)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.06)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.57)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.11)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.37)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.68)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1117
\-------------------------

Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1073; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.60)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.61)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.47)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.61)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.63)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.63)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.63)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.53)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.50)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.86)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.01)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.32)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.32)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.12)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.92)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1118
\-------------------------

Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1071; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.69)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.13)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.85)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.75)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.66)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.56)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.74)
72% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1119
\-------------------------

Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1069; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.46)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.05)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.10)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.10)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.58)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.56)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.05)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.05)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.05)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.05)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.87)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded -0.12)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded -0.12)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded -0.12)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded -0.12)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded -0.12)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 1.16)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded 1.56)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.38)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.38)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.38)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.38)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.57)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.89)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded -0.26)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 0.81)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 0.81)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 0.81)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.23)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.37)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.37)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.00)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.14)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.54)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.13)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.65)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 0.23)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1120
\-------------------------

Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1067; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.88)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.29)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.87)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.34)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.56)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.36)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.36)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.72)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.02)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.94)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.77)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.43)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1121
\-------------------------

Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1065; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.02)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.48)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.48)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.48)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.48)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.48)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.48)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.48)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.51)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.51)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.51)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.78)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.53)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.53)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.53)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.53)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.53)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.10)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.67)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.99)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.82)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.89)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.59)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.65)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.65)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.65)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.65)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.77)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.81)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.81)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.81)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.81)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.81)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.99)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1122
\-------------------------

Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1062; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 1.29)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.28)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.92)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.90)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.79)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.23)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.71)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1123
\-------------------------

Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1060; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.51)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.29)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.87)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.51)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.51)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.51)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.51)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.51)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.51)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.51)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.51)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.51)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.38)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.24)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.39)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.89)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 0.93)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.06)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.06)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.06)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.06)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.06)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.01)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.62)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.05)
47% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1124
\-------------------------

Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1058; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.52)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.92)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.77)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.57)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.68)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.70)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.40)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.40)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.40)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.99)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.52)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.91)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.77)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.08)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.08)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.08)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.51)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.43)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.46)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.51)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.03)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.75)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.80)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.65)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.10)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.10)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.10)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.24)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.26)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.20)
4% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1125
\-------------------------

Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1056; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.83)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.11)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'right')
Agent attempted driving left through a red light. (rewarded -9.28)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.47)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.86)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.69)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.38)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.61)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.61)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.61)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.61)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent followed the waypoint right. (rewarded 2.52)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.18)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.18)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.94)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.88)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.22)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 1.20)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.30)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.43)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.85)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.50)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.48)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.29)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1126
\-------------------------

Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1054; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.07)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.29)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.99)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.41)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.54)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.67)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1127
\-------------------------

Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1052; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.91)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.67)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.31)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.95)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.19)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.04)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.01)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.01)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.01)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.12)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.65)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.65)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.65)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.65)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.65)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.93)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.69)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.47)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.47)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.47)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.47)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.41)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.36)
54% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.36)
54% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.36)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.13)
51% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.13)
51% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.13)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
49% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
49% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
49% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
49% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
49% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
49% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
49% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
49% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.21)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.14)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.14)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.14)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.14)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.14)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.14)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.14)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.14)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.71)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.45)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.16)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.16)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.16)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.16)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.16)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
29% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
29% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
29% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.17)
26% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1128
\-------------------------

Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1050; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.55)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.28)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.50)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.50)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.50)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 1.74)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.87)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded 1.42)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded 1.42)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded 1.42)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded 1.42)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.42)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.91)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.04)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.04)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.04)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.81)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.36)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.64)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.92)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.45)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.75)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.90)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.54)
8% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1129
\-------------------------

Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1048; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.80)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.74)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.70)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.70)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.70)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.70)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.70)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.70)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.70)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.65)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.13)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.13)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.66)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.66)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.55)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.55)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.55)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.29)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.52)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.52)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.52)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.70)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.20)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.45)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.68)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.62)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.62)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.62)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.62)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.62)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.62)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.62)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.62)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.62)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.00)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.65)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.65)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.65)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.65)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.65)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.28)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.15)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.03)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.06)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.48)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.34)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.94)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.23)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.30)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.64)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.11)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.62)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.77)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1130
\-------------------------

Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1046; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.95)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'right')
Agent drove right instead of forward. (rewarded 1.18)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.10)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.45)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.45)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.42)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.44)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.04)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.67)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.43)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.03)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.06)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.97)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.20)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.66)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1131
\-------------------------

Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1044; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.54)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.10)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.29)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.49)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.18)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.18)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.85)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.02)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.00)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.48)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.49)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.64)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.64)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.64)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.27)
47% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1132
\-------------------------

Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1041; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.99)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 1.54)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.15)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.25)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.82)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.45)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.21)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.64)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.49)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent drove forward instead of right. (rewarded 1.32)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.21)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.21)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1133
\-------------------------

Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1039; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.92)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.84)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.58)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.82)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.82)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.82)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.66)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.67)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.67)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.67)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.86)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.17)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.70)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.44)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.44)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.44)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.44)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.25)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.25)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.25)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.25)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.25)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 0.04)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.19)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.11)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1134
\-------------------------

Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1037; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.73)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.36)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.05)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.45)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.83)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.31)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1135
\-------------------------

Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1035; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.39)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.78)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.96)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.14)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.19)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.89)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.89)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.89)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.06)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.63)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.08)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1136
\-------------------------

Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1033; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.88)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.14)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.70)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.81)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.70)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.70)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.70)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.70)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.70)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.64)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.92)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.80)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1137
\-------------------------

Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1031; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.29)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.19)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.55)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.13)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.10)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.74)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.55)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.84)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.99)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.70)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.06)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.25)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.25)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.25)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.04)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.78)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1138
\-------------------------

Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1029; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.95)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.80)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.04)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.95)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.32)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.15)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.77)
77% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1139
\-------------------------

Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1027; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -9.59)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.34)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.95)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.69)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.76)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.92)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.39)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.39)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.25)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 2.02)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.19)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1140
\-------------------------

Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1025; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.86)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.81)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.90)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.25)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.30)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.17)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.13)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.13)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.13)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.13)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 0.10)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.83)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.55)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.17)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.17)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.17)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.17)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.17)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.17)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', 'forward')
Agent properly idled at a red light. (rewarded 2.59)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 2.46)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.31)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', 'right')
Agent drove right instead of forward. (rewarded -0.16)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.07)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.36)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.71)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 0.51)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.50)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.37)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.58)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.90)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1141
\-------------------------

Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1023; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.92)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.56)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.49)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.19)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.76)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1142
\-------------------------

Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1021; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.30)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.25)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.50)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1143
\-------------------------

Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1019; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.08)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.85)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.94)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.94)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.94)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.77)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.73)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.88)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.85)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.95)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.91)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.91)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.51)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'right')
Agent followed the waypoint forward. (rewarded 1.62)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1144
\-------------------------

Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1017; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.76)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.83)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.73)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.84)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.95)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.95)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.95)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.95)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.95)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.45)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.34)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.28)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.67)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.41)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.69)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.21)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.06)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.68)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.47)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.53)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 0.90)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 0.90)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 0.90)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 0.90)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 0.90)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 0.63)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.45)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.38)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1145
\-------------------------

Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1015; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.85)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.48)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.75)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.96)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.94)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.94)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.94)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.86)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.04)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.04)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.04)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.39)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.41)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.27)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.00)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.31)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.24)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.39)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.94)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.63)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.51)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1146
\-------------------------

Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1013; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.97)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.01)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.15)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.79)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 0.53)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.08)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1147
\-------------------------

Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1011; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 3.00)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.65)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.42)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.42)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.42)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.42)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 0.95)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.66)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.66)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.66)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.66)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.95)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.25)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.25)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.13)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.13)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.13)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.46)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.65)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.90)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1148
\-------------------------

Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1009; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.98)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.27)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.27)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.27)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.27)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.27)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.27)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.66)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.08)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.82)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.82)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.82)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.82)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.82)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.82)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.82)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.23)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.90)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.90)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.90)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.90)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.90)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.90)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.90)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.90)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.90)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.90)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.90)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.02)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.13)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.98)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1149
\-------------------------

Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1007; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.08)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.98)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.83)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded 1.45)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.97)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.70)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.57)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.57)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.57)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.57)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.57)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.57)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.76)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.89)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.58)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.58)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.58)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.12)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.76)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.76)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.76)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.76)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.76)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.75)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.30)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.41)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1150
\-------------------------

Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1005; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.70)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.93)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.83)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.41)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.41)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.41)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.22)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.22)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.22)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.22)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.22)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.92)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1151
\-------------------------

Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1003; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.50)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.75)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.58)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.47)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.97)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.97)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.97)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.97)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.97)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.97)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.97)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.97)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.55)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.55)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.92)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.92)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.92)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.92)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.84)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.44)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.52)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.29)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.58)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.57)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.57)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.57)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.57)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.57)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.91)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.49)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1152
\-------------------------

Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.1001; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.39)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.02)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.36)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.68)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.68)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.52)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.70)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.28)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.28)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.28)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.28)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.67)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.83)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.83)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.43)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.06)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.74)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.74)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.85)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.85)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', 'right')
Agent drove forward instead of right. (rewarded 0.01)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.59)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.59)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.53)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.76)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.29)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.87)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1153
\-------------------------

Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0999; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.36)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.90)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.96)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.97)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.25)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded -0.08)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.64)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.59)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.89)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.45)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.52)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1154
\-------------------------

Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0997; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.98)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 1.47)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.80)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.81)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.08)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.49)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.44)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.40)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.40)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.40)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.40)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.90)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1155
\-------------------------

Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0995; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'left')
Agent drove forward instead of left. (rewarded 1.42)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.93)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.44)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.96)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.53)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.15)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.09)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.97)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.33)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.59)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.59)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.59)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.59)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.59)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.03)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.93)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.35)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.35)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.35)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.71)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.12)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1156
\-------------------------

Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0993; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.88)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.20)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.98)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.98)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.98)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.98)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.98)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.98)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.99)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.69)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.69)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.69)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.69)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.61)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1157
\-------------------------

Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0991; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.25)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.22)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.22)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.22)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.22)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.22)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.22)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.43)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.44)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.67)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.90)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.90)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.90)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.90)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.90)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'right')
Agent drove forward instead of right. (rewarded 1.46)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.08)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.15)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.15)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.18)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.91)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.91)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.20)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.44)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1158
\-------------------------

Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0989; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.65)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.05)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.06)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.64)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.91)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.91)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.91)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.56)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.65)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.56)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1159
\-------------------------

Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0987; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.68)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.59)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.69)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.14)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.50)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.88)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.43)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.39)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.39)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.39)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.56)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.71)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent drove right instead of left. (rewarded 0.41)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent drove right instead of left. (rewarded 0.41)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent drove right instead of left. (rewarded 0.41)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 0.95)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.60)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent drove right instead of left. (rewarded 0.29)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.11)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.08)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.00)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.72)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.79)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1160
\-------------------------

Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0985; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 1.12)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.29)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.29)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.29)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.29)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.29)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.29)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.62)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.32)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.32)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.32)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.32)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.32)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.07)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.46)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.83)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.48)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.22)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.45)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1161
\-------------------------

Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0983; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.96)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.24)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.93)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.93)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.93)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.93)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.50)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.50)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.20)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.81)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.21)
72% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1162
\-------------------------

Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0981; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.05)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.75)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.62)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 2.23)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.38)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.72)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.15)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.37)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.29)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.29)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.68)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.04)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1163
\-------------------------

Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0979; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.28)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.94)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.44)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.95)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.85)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.46)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.46)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.23)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.17)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.17)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.17)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.36)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.71)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.43)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.21)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.21)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.21)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.91)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.23)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.24)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.24)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.24)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.24)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.24)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.99)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.48)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.86)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1164
\-------------------------

Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0977; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.41)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.31)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.90)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.99)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.42)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.06)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.41)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.19)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.62)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.62)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.62)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.62)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.62)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.62)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 1.59)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.04)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.04)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.48)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.48)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.48)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.60)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.57)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.14)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.25)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.25)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.25)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.25)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.59)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1165
\-------------------------

Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0975; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.65)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.86)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.29)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.62)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.21)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.21)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.21)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.18)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1166
\-------------------------

Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0973; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.08)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.86)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.61)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.28)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.75)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.71)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.69)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.42)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.42)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.42)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.42)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.76)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.83)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.83)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.83)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.71)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.71)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.88)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.75)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.13)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.13)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.19)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.19)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.19)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.58)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.88)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.03)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.98)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.82)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.69)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.03)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 0.90)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.20)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.88)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.12)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1167
\-------------------------

Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0971; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.09)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.73)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.08)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.00)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.09)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.67)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.41)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.73)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.42)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.71)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.24)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.70)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.66)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.06)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.73)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.25)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.74)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.20)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.61)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.38)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.47)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.78)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.78)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.94)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.28)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.30)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.30)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.30)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.30)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.69)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.69)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.69)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 0.50)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 0.50)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 0.50)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.52)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.23)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1168
\-------------------------

Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0969; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.13)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.79)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.79)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.79)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.79)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.29)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.85)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.54)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.13)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.91)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.53)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.87)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.62)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.73)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.53)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.83)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.44)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.72)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.93)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.04)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.75)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.50)
27% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1169
\-------------------------

Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0967; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.41)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.88)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.95)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.46)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.40)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.40)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.40)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.68)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.49)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.59)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.86)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.86)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.86)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.40)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.99)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.78)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.10)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.07)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1170
\-------------------------

Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0965; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.17)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.00)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.29)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.29)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.29)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.79)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.12)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.12)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.12)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.12)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove forward instead of left. (rewarded 0.81)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.27)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.28)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1171
\-------------------------

Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0963; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.66)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.86)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'right')
Agent followed the waypoint right. (rewarded 1.46)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.98)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.98)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.98)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.98)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.56)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.16)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.79)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.79)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.79)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.79)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.05)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.51)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1172
\-------------------------

Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0961; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.99)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.59)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.29)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.25)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.13)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 0.97)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1173
\-------------------------

Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0959; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.69)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.78)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.37)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.37)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.37)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.81)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.17)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.08)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.55)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.55)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.55)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.55)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.67)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.84)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.72)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.36)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.13)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.03)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.06)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.97)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.88)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1174
\-------------------------

Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0958; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.09)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.14)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.32)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.04)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.54)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.54)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.54)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.54)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.54)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.05)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.34)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.31)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.88)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.74)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.80)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.92)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.19)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.82)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.47)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.47)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.47)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.47)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.47)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.31)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1175
\-------------------------

Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0956; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.47)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.60)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.09)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.09)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.09)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.09)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.09)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.09)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.09)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.09)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.57)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.13)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.89)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.18)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.29)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.64)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.62)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.63)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.73)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.73)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.73)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.73)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.73)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 0.90)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1176
\-------------------------

Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0954; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 2.66)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.22)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.66)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.62)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.62)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.08)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.50)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.84)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.05)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.05)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.05)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.05)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.05)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1177
\-------------------------

Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0952; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.90)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.79)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.88)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 0.61)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.04)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.04)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.58)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.92)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded -0.04)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded -0.04)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded -0.04)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded -0.04)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded -0.04)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.88)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.43)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.43)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.43)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.43)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.09)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.09)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.09)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.96)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.95)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.91)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.12)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1178
\-------------------------

Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0950; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.91)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 1.13)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.72)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.71)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.22)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.27)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1179
\-------------------------

Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0948; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.59)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.39)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.05)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.57)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.35)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.77)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.77)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.42)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.15)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.54)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.21)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.13)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.13)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.13)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.13)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.13)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.31)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.54)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1180
\-------------------------

Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0946; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.74)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.18)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.85)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.85)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.85)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.85)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.85)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.90)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.40)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.03)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.55)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.15)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.79)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 0.81)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.75)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.75)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.65)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.65)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.93)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.93)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.93)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.93)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.93)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.93)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.11)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.54)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1181
\-------------------------

Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0944; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.81)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.53)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.14)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.39)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.44)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.46)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.46)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.78)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1182
\-------------------------

Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0942; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.77)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.59)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.89)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.97)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.97)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.97)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.97)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.11)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.65)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.69)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.94)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.02)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.44)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.81)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1183
\-------------------------

Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0940; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.62)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.67)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.16)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.87)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.92)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.81)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.81)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.80)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.15)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.15)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.15)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.15)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.57)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.61)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.61)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1184
\-------------------------

Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0939; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.06)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.42)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.31)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.88)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.72)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.63)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.63)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1185
\-------------------------

Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0937; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.43)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.98)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.21)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.21)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.21)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.65)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.16)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.68)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.64)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.85)
63% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1186
\-------------------------

Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0935; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.75)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.93)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.57)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.03)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.54)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.54)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.54)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.54)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.28)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.60)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.41)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.30)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.56)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.62)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.68)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.50)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.67)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.67)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.67)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'right')
Agent drove forward instead of left. (rewarded 0.30)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.76)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.14)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded -0.55)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.04)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.37)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1187
\-------------------------

Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0933; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.78)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.10)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.13)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.13)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.06)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.01)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.68)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.68)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.68)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.68)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.08)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.08)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.73)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.73)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.73)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.73)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.02)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.94)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.65)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1188
\-------------------------

Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0931; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.54)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.13)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.28)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.20)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.02)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.61)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.54)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.54)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.54)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.54)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.54)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.54)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.55)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.14)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.33)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.25)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.92)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.36)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1189
\-------------------------

Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0929; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.88)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.13)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.37)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.40)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.40)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.40)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.05)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.05)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.31)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.31)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.64)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.57)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.57)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.57)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.57)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.43)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.80)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.80)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.11)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.11)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.79)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.94)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.46)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.46)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.93)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.74)
33% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.74)
33% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.74)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.79)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.79)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.79)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 1.45)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.42)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.68)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.59)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.59)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.59)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.59)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.11)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.37)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 0.81)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.43)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.16)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1190
\-------------------------

Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0927; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.32)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.46)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.98)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.30)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.46)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.51)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.82)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.82)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.82)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.00)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.22)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.73)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.73)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.25)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1191
\-------------------------

Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0926; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.55)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.00)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.36)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 2.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 2.62)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.30)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1192
\-------------------------

Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0924; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.43)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.39)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.10)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.10)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.10)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.10)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.10)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.10)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.03)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.86)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.93)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.88)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.71)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1193
\-------------------------

Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0922; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.85)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.65)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.45)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.05)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.69)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.88)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.24)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.46)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.95)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.95)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.95)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.95)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.97)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.63)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.27)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.83)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.07)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.05)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1194
\-------------------------

Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0920; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.23)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.10)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.80)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.10)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.10)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.10)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.20)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'right', 'left')
Agent drove right instead of left. (rewarded 0.28)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'right', 'left')
Agent drove right instead of left. (rewarded 0.28)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.53)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.53)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.03)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.03)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.03)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.70)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.51)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.51)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.64)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1195
\-------------------------

Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0918; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.25)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.41)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.00)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.11)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.95)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.99)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.74)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.82)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.08)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.40)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.22)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.22)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.09)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.06)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.91)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.91)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.91)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.91)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.52)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.98)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.98)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.98)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 1.46)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.75)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.98)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.17)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.24)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.12)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1196
\-------------------------

Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0916; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove forward instead of left. (rewarded 0.65)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.18)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.45)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.24)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.66)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.34)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.08)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.57)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.57)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.57)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.57)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.57)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.57)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.33)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.20)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.03)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.90)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1197
\-------------------------

Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0914; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.38)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.10)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.11)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.11)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.01)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.72)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.48)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.84)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1198
\-------------------------

Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0913; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.85)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent followed the waypoint right. (rewarded 1.77)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.85)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.41)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.39)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.39)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.39)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.05)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1199
\-------------------------

Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0911; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.27)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.57)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.11)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.85)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.05)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.94)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.82)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.85)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.17)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.62)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.83)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.33)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.72)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.72)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.72)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.72)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.72)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.69)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.77)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.99)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.87)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1200
\-------------------------

Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0909; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.80)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.06)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.03)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.71)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.71)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.71)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.71)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.71)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.71)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.81)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.86)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.86)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.77)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.77)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.77)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.02)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.81)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.81)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.81)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.81)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.62)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1201
\-------------------------

Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0907; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.76)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.97)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.27)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.90)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.35)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 2.67)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.36)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.22)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1202
\-------------------------

Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0905; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.40)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.44)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.64)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.85)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.85)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.85)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.43)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.56)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.64)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.23)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.28)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.68)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.62)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.09)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.82)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.82)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.78)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.13)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.94)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.11)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.83)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.48)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.81)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.36)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.13)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.83)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1203
\-------------------------

Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0904; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.10)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.70)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.06)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.55)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.55)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.24)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.10)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.10)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.10)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.10)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.10)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.86)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.76)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.76)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.76)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.76)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.57)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.91)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.91)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.42)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.31)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.00)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 1.43)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 1.43)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 1.43)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.10)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.26)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.39)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.97)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.92)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.92)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.92)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.92)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.92)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.92)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.92)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.17)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1204
\-------------------------

Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0902; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.97)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.48)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.52)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.96)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.96)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.55)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.29)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.29)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.85)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.37)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.80)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.80)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.37)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.97)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.97)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.97)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.97)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.97)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.78)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1205
\-------------------------

Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0900; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.47)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.12)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.29)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.89)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.59)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.48)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.86)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.10)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.10)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.46)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.21)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.04)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.04)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.04)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.04)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.40)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.41)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.88)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.30)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.81)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.60)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.73)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.65)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.50)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.39)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.59)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.59)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.59)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.59)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.81)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.67)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.67)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.67)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.48)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.78)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.87)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1206
\-------------------------

Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0898; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.42)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.28)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -6.00)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.72)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.72)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.72)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.59)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.59)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.06)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.06)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.73)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.57)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.61)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.94)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.52)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.10)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded -0.51)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.57)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.41)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1207
\-------------------------

Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0896; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.99)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.05)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.89)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.97)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.59)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.59)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.59)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.59)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.59)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.59)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.27)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.60)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.47)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.47)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.47)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.47)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.47)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.47)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.47)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.81)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.32)
67% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1208
\-------------------------

Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0895; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.77)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.79)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.15)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.15)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.15)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.15)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.15)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.15)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.04)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.65)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'left')
Agent drove forward instead of left. (rewarded 1.42)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.52)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.21)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1209
\-------------------------

Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0893; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.22)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.82)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.51)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.43)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.43)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.43)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.25)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.42)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.61)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.50)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.50)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.50)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.73)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.72)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.72)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.65)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.65)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.65)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.65)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.46)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.51)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.51)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.26)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.26)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.26)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.26)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.26)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.18)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 1.27)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 1.27)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 1.27)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.24)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.41)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.12)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.08)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 0.89)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.13)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.46)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.46)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.46)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.46)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.46)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.15)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 0.23)
3% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded -0.31)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1210
\-------------------------

Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0891; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.26)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.51)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.16)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.58)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.82)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.82)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.82)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.82)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.64)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.71)
77% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1211
\-------------------------

Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0889; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.42)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.77)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.77)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.77)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.16)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.37)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.24)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.19)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.63)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.63)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.10)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.10)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.10)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.80)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1212
\-------------------------

Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0887; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.26)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.82)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.35)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.17)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.60)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.60)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.60)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.40)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.33)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1213
\-------------------------

Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0886; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.13)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.59)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.69)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.15)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.21)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.71)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'right')
Agent drove right instead of forward. (rewarded 0.69)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.99)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.10)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.85)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.85)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.85)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.24)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.42)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.99)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.73)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.73)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.73)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.73)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.13)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.10)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.10)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.10)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.10)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.37)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.65)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.01)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.35)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1214
\-------------------------

Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0884; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.48)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.83)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.10)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.36)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.83)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.89)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.43)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.56)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.70)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.70)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.70)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.96)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.96)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.26)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.26)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.62)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.69)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.03)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.31)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.19)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.19)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.19)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.19)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.56)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded -0.09)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded -0.09)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded -0.09)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded -0.09)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 0.52)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 0.95)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 0.95)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 0.95)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 0.95)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 0.95)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 0.95)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 0.95)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.54)
8% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1215
\-------------------------

Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0882; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.10)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.18)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.30)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.31)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.76)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.05)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.40)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.40)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.55)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.55)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.55)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.55)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.50)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.75)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded -0.29)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 1.45)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.69)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1216
\-------------------------

Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0880; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.59)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.96)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.48)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.17)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.86)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.71)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.25)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1217
\-------------------------

Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0879; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.30)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.93)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.53)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.83)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.10)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.22)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.98)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.80)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.78)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.39)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1218
\-------------------------

Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0877; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.69)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.65)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.99)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.09)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.17)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1219
\-------------------------

Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0875; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'right')
Agent followed the waypoint right. (rewarded 1.65)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.18)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.34)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.34)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.34)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.34)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.34)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.34)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.34)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.97)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.19)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.19)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.19)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.19)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.19)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.38)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.26)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.92)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.56)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.48)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.59)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.07)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1220
\-------------------------

Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0873; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.38)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.92)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.68)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.68)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.68)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.68)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.68)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.68)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.68)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.68)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.28)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.13)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.53)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.99)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.22)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded -0.04)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.63)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.98)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.77)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.97)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 0.74)
49% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 0.74)
49% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 0.74)
49% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 0.74)
49% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 0.74)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.43)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.96)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent drove forward instead of right. (rewarded 1.08)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.59)
37% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1221
\-------------------------

Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0872; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.09)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.97)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.91)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.84)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.65)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.54)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.30)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.23)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.10)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.68)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.70)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.81)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.86)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.35)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.69)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
33% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
33% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
33% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
33% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
33% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.99)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.25)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.26)
23% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1222
\-------------------------

Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0870; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.27)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.88)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.45)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.45)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.45)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.95)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.55)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.96)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.48)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.78)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1223
\-------------------------

Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0868; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.88)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.91)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.54)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.18)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.81)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.34)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.12)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.22)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.59)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.22)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.23)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.84)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.40)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.40)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.40)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.40)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.40)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.26)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.53)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.44)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.93)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.56)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.56)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.56)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.56)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.56)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.56)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.56)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.56)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.56)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', 'forward')
Agent properly idled at a red light. (rewarded 2.01)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.56)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1224
\-------------------------

Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0866; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'right')
Agent drove right instead of forward. (rewarded 1.25)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.35)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.01)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.26)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.79)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.12)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.15)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.24)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.02)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.42)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.42)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.42)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.42)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.42)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 1.58)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.61)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.61)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.21)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.21)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.21)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.21)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.21)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.21)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.21)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.32)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.78)
54% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.78)
54% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.78)
54% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.78)
54% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.78)
54% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.78)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.89)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.59)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.94)
46% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.94)
46% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.94)
46% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.94)
46% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.94)
46% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.94)
46% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.94)
46% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.94)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.12)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.50)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded -0.31)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.78)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded -0.24)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.34)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.75)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.51)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.16)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.86)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.86)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.86)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.86)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.86)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.86)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.86)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.88)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.85)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded -0.13)
9% of time remaining to reach destination.

/-------------------
| Step 32 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.77)
6% of time remaining to reach destination.

/-------------------
| Step 33 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.23)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.23)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.23)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.23)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.23)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.23)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.23)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.23)
3% of time remaining to reach destination.

/-------------------
| Step 34 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.61)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1225
\-------------------------

Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0865; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent drove forward instead of left. (rewarded 0.90)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.59)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.81)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.93)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.93)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.93)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.93)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.98)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.58)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.08)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.18)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.56)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.56)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.56)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.56)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.56)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.56)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.56)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.56)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.12)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.58)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.23)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.23)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.23)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.23)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.23)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.23)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.23)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.23)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.23)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.83)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.99)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1226
\-------------------------

Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0863; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.40)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.42)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.74)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.87)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.87)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.87)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.38)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.26)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.13)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 1.25)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.25)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.25)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.25)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.62)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.54)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.58)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.47)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.13)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.43)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 0.92)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.33)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.60)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded -0.28)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.52)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.52)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.52)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.52)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.52)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.52)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.52)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.80)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1227
\-------------------------

Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0861; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 1.97)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.60)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.15)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.55)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.95)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.60)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.60)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.52)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.44)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.69)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.55)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1228
\-------------------------

Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0859; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.63)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.31)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.39)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'right')
Agent followed the waypoint forward. (rewarded 2.74)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.68)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.87)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.80)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.11)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.11)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.11)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.37)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.27)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1229
\-------------------------

Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0858; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.87)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.07)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.85)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.57)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.59)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.59)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.59)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.59)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.25)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.91)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.50)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.30)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.30)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.80)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.97)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent drove forward instead of right. (rewarded -0.22)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.04)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.04)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.04)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.04)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.97)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.18)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.54)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.56)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1230
\-------------------------

Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0856; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.44)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.39)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.21)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.79)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.03)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.91)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.06)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1231
\-------------------------

Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0854; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.84)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.77)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.53)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.92)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.05)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.05)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.08)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.08)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.08)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.08)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.08)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.55)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.08)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.97)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1232
\-------------------------

Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0853; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.08)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.75)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.87)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.87)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.24)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.18)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1233
\-------------------------

Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0851; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.85)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.96)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.97)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.38)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.92)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.78)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.22)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.47)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 0.21)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 0.21)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.27)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.32)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1234
\-------------------------

Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0849; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.40)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.66)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.11)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.36)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.32)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.14)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.54)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.68)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1235
\-------------------------

Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0848; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.59)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.17)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.54)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.34)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.68)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.16)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.03)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.51)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'right')
Agent drove left instead of forward. (rewarded 1.59)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.33)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.41)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.86)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.86)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.86)
57% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.86)
57% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.86)
57% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.86)
57% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.86)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.11)
51% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.11)
51% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.11)
51% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.11)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.91)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.37)
43% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1236
\-------------------------

Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0846; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.08)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.02)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.74)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.85)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.97)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.30)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.74)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.74)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.74)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.74)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.74)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.74)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.74)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.20)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1237
\-------------------------

Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0844; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.66)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.69)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.30)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.87)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.85)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.13)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.03)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.03)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.03)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.35)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1238
\-------------------------

Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0842; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.62)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.65)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.84)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.19)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.30)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.36)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.69)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.99)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1239
\-------------------------

Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0841; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.08)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.30)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.67)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.67)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.67)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.67)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.67)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.55)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.97)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.97)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.97)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.97)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.20)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.11)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.63)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1240
\-------------------------

Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0839; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.89)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.78)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.57)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.10)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.10)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.10)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.10)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.10)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.10)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.10)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.10)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.10)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.70)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1241
\-------------------------

Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0837; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.87)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.60)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.01)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.01)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.01)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.01)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.39)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.26)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.31)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.20)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.49)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.89)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.79)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.67)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.47)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.40)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.71)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1242
\-------------------------

Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0836; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.00)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.47)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.87)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.87)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.87)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.87)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.43)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.72)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.43)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.43)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.11)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.84)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.42)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.69)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.45)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.47)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.11)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.15)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.83)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.42)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'right')
Agent followed the waypoint right. (rewarded 1.66)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.34)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.42)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.87)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.87)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.21)
4% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1243
\-------------------------

Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0834; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.56)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.87)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.84)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.68)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1244
\-------------------------

Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0832; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.91)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 2.80)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.15)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.85)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.16)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.51)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.13)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.85)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.85)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.85)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.85)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.91)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.81)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.81)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.32)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.62)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.23)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.32)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.83)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.09)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1245
\-------------------------

Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0831; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.13)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.08)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.30)
91% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.30)
91% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.30)
91% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.30)
91% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.30)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.20)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.20)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.20)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.20)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.11)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.52)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.60)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.99)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.66)
71% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1246
\-------------------------

Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0829; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.11)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.41)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.21)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.41)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 1.65)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.91)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.91)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.91)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.79)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.79)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.29)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1247
\-------------------------

Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0827; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.42)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.83)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.94)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.94)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.94)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.94)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.94)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.12)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.07)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.00)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.82)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.32)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.82)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.56)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.77)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1248
\-------------------------

Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0826; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.90)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.64)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.33)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.33)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.33)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.71)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.49)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.70)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.70)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.70)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.70)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.70)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.69)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.98)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.51)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.55)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.43)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.47)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.47)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.25)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.81)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1249
\-------------------------

Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0824; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.80)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 0.15)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.17)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.53)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.34)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.34)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.34)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.34)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.77)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.40)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.02)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.00)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1250
\-------------------------

Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0822; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.08)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.42)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.47)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.48)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.25)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.25)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.25)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.25)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.25)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.31)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.41)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.41)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.50)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.88)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.55)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 0.50)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.66)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.38)
48% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.38)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.25)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.25)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.22)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.22)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.70)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1251
\-------------------------

Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0821; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.02)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.61)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.78)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.91)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.98)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.90)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.63)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.63)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.63)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.96)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.79)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1252
\-------------------------

Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0819; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.76)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.22)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.35)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.86)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.48)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.81)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.77)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.03)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.20)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.20)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.20)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.20)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.78)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.53)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.31)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 0.00)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 1.04)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 1.04)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.14)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.14)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.14)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.14)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.50)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.62)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.62)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.62)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.57)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.57)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 0.62)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.72)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.95)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1253
\-------------------------

Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0818; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'right')
Agent drove forward instead of left. (rewarded 0.70)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.93)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.14)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.31)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.04)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.61)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.61)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.61)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.61)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.50)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.16)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.01)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.46)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.76)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'right')
Agent followed the waypoint right. (rewarded 2.69)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.76)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.06)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.68)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.24)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.92)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.34)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1254
\-------------------------

Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0816; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.89)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.83)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.59)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.25)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.68)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.19)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.98)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.31)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.20)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.20)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.35)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.30)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.30)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.30)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.15)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.61)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.61)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.61)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.61)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.61)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.56)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.71)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.85)
49% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.85)
49% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.85)
49% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.85)
49% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.85)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
46% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1255
\-------------------------

Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0814; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.63)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.23)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.93)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.83)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.81)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.81)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.28)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.28)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.28)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.28)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.22)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 0.98)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 0.98)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 0.98)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.66)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.54)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.17)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.69)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.69)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.69)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.69)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.37)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.55)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.29)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1256
\-------------------------

Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0813; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.41)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.21)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.08)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.22)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.14)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.72)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.16)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.40)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.40)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.40)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.40)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.40)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.40)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.40)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.40)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.25)
63% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1257
\-------------------------

Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0811; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.12)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.69)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.78)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.78)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.78)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.78)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.01)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.29)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.37)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.07)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.95)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.14)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.14)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.14)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.92)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.92)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.92)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.72)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1258
\-------------------------

Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0809; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.46)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.44)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.52)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.96)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.96)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.48)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.48)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.48)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.48)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.37)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.80)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.80)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.79)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.98)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.16)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.70)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.70)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.12)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1259
\-------------------------

Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0808; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.20)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.98)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.32)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.32)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.29)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.60)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.15)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1260
\-------------------------

Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0806; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.73)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.25)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.48)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.48)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.16)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.38)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.71)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.46)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.70)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.93)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.46)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.84)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.30)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.02)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.59)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.50)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.18)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.18)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.18)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.18)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.18)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.04)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.57)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.57)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.57)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.57)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.57)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.31)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1261
\-------------------------

Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0805; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.79)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.68)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.39)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.66)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.99)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.99)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.99)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.75)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.26)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.60)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.82)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.54)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.94)
57% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1262
\-------------------------

Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0803; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.40)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.53)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.10)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.53)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.65)
86% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.65)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.76)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.76)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.76)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.76)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.84)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.50)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.50)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.50)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.50)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.50)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.50)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.41)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.69)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.54)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.04)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.60)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.74)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.81)
49% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.81)
49% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.81)
49% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.81)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.99)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.46)
43% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1263
\-------------------------

Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0801; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.88)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.10)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.61)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.61)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.45)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.69)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1264
\-------------------------

Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0800; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.76)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.33)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.78)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.63)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.72)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.10)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.39)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.07)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 1.63)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 1.63)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 1.63)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 1.63)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.89)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.76)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.68)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.68)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.68)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.68)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.68)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.16)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.80)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.80)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.80)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.80)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.80)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.80)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.80)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.15)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.82)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.51)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1265
\-------------------------

Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0798; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'right')
Agent drove right instead of forward. (rewarded 1.83)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.27)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.23)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.48)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.19)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.19)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.19)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.19)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.19)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.19)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.19)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.49)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.42)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.73)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.12)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.12)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.12)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.12)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.12)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.12)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.12)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.12)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.12)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.65)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.79)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.79)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.79)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.79)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.79)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.79)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.27)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.04)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.16)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.46)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 0.88)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 0.88)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 0.88)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 0.88)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.83)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.39)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.80)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.87)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1266
\-------------------------

Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0797; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 2.34)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.50)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.09)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.56)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.46)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.66)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.97)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1267
\-------------------------

Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0795; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.44)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.23)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.26)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.67)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.67)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.60)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.58)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.58)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.58)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.58)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.58)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.04)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.47)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.64)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.64)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.64)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.64)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.64)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.26)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.66)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.83)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.83)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.83)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.27)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.27)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.21)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.97)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.97)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.76)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.43)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.10)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.40)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1268
\-------------------------

Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0793; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.73)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.49)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.43)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.43)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.43)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.40)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 0.94)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.50)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.03)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.29)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.29)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.29)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.48)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.04)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.04)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.04)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.04)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.18)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.18)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.56)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.56)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.56)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.13)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.04)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1269
\-------------------------

Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0792; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.94)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.54)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.65)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.99)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.01)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.74)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.74)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.74)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.74)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.75)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.30)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.57)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.47)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.75)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.91)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.36)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.15)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.30)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.30)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.30)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.13)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.12)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.12)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.12)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.12)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.34)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 0.68)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.74)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.89)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1270
\-------------------------

Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0790; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.52)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.83)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.75)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.30)
89% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.30)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.07)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.07)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.59)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.59)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.59)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.59)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.31)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.64)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.64)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.75)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'right')
Agent drove forward instead of right. (rewarded 0.26)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.17)
66% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1271
\-------------------------

Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0789; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.16)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.21)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.31)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.23)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.89)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.71)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.14)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.35)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.77)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1272
\-------------------------

Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0787; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.58)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.94)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.88)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.94)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.82)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.24)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.24)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.24)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.39)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.53)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.35)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.04)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.04)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.04)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.04)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.04)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.66)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.30)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.30)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.30)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.91)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.91)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.43)
36% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1273
\-------------------------

Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0786; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.35)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.40)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'forward')
Agent followed the waypoint forward. (rewarded 2.22)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.39)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1274
\-------------------------

Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0784; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.48)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.26)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.61)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.95)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.22)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.03)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.03)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.03)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.15)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.26)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.86)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1275
\-------------------------

Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0782; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.21)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.36)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.65)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.98)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.61)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.80)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.43)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.40)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.01)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.34)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.56)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.37)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.35)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.86)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1276
\-------------------------

Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0781; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.16)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.00)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.00)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.01)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.61)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.01)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.01)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.06)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.70)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.97)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.73)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.81)
63% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1277
\-------------------------

Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0779; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.50)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.71)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.72)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.72)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.72)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.05)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.05)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.05)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.05)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.05)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.46)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.52)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.46)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.74)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.74)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.74)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.74)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.74)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.44)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.60)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.60)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.14)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.00)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.07)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 1.48)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.99)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.99)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.99)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.99)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.75)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.43)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.07)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.44)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.44)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.44)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.29)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.36)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.36)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.36)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.36)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.36)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.18)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.75)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.43)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1278
\-------------------------

Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0778; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.10)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.37)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.31)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.05)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.05)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.05)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.05)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.59)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.84)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.96)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.96)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.96)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.96)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.96)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.96)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.96)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.96)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.96)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.96)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.96)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.96)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.96)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.96)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.96)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.10)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.02)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.02)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.02)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.02)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.02)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.02)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.02)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.33)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.77)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.77)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.77)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.77)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.77)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.77)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.77)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.71)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1279
\-------------------------

Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0776; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.07)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.79)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.31)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.81)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.60)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.66)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.66)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.66)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.66)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.66)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.66)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.66)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.66)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.66)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.69)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.51)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 1.29)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.69)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.13)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 0.94)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.76)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.76)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.76)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1280
\-------------------------

Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0775; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.84)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.98)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.58)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.38)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.20)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.09)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1281
\-------------------------

Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0773; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.34)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.41)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.67)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.04)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.61)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.61)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.61)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.28)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.00)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.43)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1282
\-------------------------

Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0772; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.99)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.61)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.92)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.92)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.45)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.45)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.45)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.45)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.22)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.39)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1283
\-------------------------

Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0770; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.25)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.57)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.89)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 0.48)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.46)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.08)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.08)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.08)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.08)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.08)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.33)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.63)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.06)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.59)
63% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.59)
63% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.59)
63% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.59)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.21)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.15)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.52)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.52)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.52)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.28)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.46)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.46)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.46)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.01)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.78)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.85)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.85)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.85)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.85)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.76)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.97)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.06)
23% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1284
\-------------------------

Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0768; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.40)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.55)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.88)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.88)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.88)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.59)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.08)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.23)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1285
\-------------------------

Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0767; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'right', None)
Agent drove right instead of left. (rewarded 0.95)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.57)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.96)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.28)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.23)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.94)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.92)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.91)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.33)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.97)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.62)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.62)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.62)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.62)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.62)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.18)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.37)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.55)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.59)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.25)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.43)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.43)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.43)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.76)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1286
\-------------------------

Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0765; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.49)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.67)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.97)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.80)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.89)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.11)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.06)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.77)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.77)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.77)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.77)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.71)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.58)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.15)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.25)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.77)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.71)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.02)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.99)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.83)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.75)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.60)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1287
\-------------------------

Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0764; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.90)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 0.98)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.93)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.90)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.18)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.36)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.14)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1288
\-------------------------

Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0762; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.26)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.84)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.12)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.12)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.12)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.12)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.12)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.12)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.12)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.33)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1289
\-------------------------

Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0761; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.99)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.93)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.10)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.56)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.45)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.45)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.45)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.01)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.74)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.21)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.59)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.59)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.59)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.59)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 0.89)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.49)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1290
\-------------------------

Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0759; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.40)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.06)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.75)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.15)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.15)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.69)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.69)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.69)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.69)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.45)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.45)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.33)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.74)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.99)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1291
\-------------------------

Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0758; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 1.33)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.72)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.74)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.74)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.74)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.74)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.74)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.74)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.94)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.51)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 1.28)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.08)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.83)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.90)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.70)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.70)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.70)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.70)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.41)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.75)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1292
\-------------------------

Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0756; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.28)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.98)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.39)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.74)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.35)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.24)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.74)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.91)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.91)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.91)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.54)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1293
\-------------------------

Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0755; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.36)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.35)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.35)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.35)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.35)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.35)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.81)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.95)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.14)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.82)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.13)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.92)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.37)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.25)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.25)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.25)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.25)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.25)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.59)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.50)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.60)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.88)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1294
\-------------------------

Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0753; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.68)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.71)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.32)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.63)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.12)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.77)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.36)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.20)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.20)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.95)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.73)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.58)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.24)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1295
\-------------------------

Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0752; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.94)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.03)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.89)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.27)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.70)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.52)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1296
\-------------------------

Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0750; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.66)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.87)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.68)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.84)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.84)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.56)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.04)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.20)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.75)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1297
\-------------------------

Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0749; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.69)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.69)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.69)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.69)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.69)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.69)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.69)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.69)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.94)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1298
\-------------------------

Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0747; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.62)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.29)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.82)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.11)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.05)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.05)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.05)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.05)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.55)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1299
\-------------------------

Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0746; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.19)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.43)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.78)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.72)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.31)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.39)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.80)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.04)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.04)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.06)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1300
\-------------------------

Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0744; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.02)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.53)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.69)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.63)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.63)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.40)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.63)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.56)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.65)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.65)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.65)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.65)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.74)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.42)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1301
\-------------------------

Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0743; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.01)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.03)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.16)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.22)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.41)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.69)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.33)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.37)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.56)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.37)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded -0.05)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent followed the waypoint right. (rewarded 1.80)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.24)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.75)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.17)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.72)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.72)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.72)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.72)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.72)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.25)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.31)
16% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1302
\-------------------------

Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0741; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.54)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.19)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.12)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.12)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.12)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.12)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.06)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.24)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.90)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.18)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.07)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.10)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.10)
69% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1303
\-------------------------

Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0740; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.02)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.42)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.59)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.86)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.60)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.60)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.12)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.99)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.48)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.48)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.56)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.35)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.00)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.59)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1304
\-------------------------

Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0738; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.70)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.22)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.22)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.22)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.22)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.22)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.22)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.22)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.66)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.31)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.73)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.40)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.88)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.64)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.64)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.64)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.64)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.64)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.76)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.12)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.12)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.13)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.41)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.48)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.70)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.70)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.70)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.70)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.70)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1305
\-------------------------

Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0737; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.52)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', 'forward')
Agent properly idled at a red light. (rewarded 2.67)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.90)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.21)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.29)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.76)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.77)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.53)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.80)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.09)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.09)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.27)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.27)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.27)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.27)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.47)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.11)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.11)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.11)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.62)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.57)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.57)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.57)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.57)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.57)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.57)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.57)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.57)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.57)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.57)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.21)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1306
\-------------------------

Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0735; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.51)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.69)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.51)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.24)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.35)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.42)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.29)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.91)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.56)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1307
\-------------------------

Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0734; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.04)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.43)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.03)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.28)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.31)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.02)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.47)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.61)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1308
\-------------------------

Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0732; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.23)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.12)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.43)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.33)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.42)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.42)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.42)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.42)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.42)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.38)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1309
\-------------------------

Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0731; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.78)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.72)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.78)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.21)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.77)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.77)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.77)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.43)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1310
\-------------------------

Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0729; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.58)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.18)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.10)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.59)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.01)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.84)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.84)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.16)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.25)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.32)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.57)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.98)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.06)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.08)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.08)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.39)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.14)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.57)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.57)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded -0.72)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.39)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1311
\-------------------------

Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0728; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.68)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.28)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.90)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.90)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.90)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.90)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.27)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.05)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1312
\-------------------------

Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0727; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.33)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.62)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.54)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.54)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.54)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.54)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.98)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.16)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.16)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.16)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.31)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.45)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1313
\-------------------------

Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0725; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.30)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.85)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.14)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.64)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1314
\-------------------------

Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0724; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.66)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.44)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.83)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.47)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.19)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.70)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1315
\-------------------------

Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0722; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.18)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.39)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.39)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1316
\-------------------------

Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0721; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.78)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.73)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.23)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.23)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.35)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.03)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.03)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.03)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.62)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.79)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.19)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.80)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.69)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.19)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.08)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded -0.05)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.44)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.98)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.88)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.34)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.93)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.34)
37% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1317
\-------------------------

Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0719; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.26)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.90)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.08)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.46)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.65)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.65)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.32)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.96)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.54)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.31)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.06)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.34)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.83)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1318
\-------------------------

Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0718; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.78)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.15)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.59)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.59)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.59)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.59)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.59)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.59)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.59)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.59)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.51)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.26)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.88)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.23)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.27)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.79)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.51)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1319
\-------------------------

Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0716; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.92)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.33)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.33)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.05)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.48)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1320
\-------------------------

Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0715; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.77)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.25)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.07)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.33)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.39)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.39)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1321
\-------------------------

Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0714; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.32)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.54)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.93)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.79)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.37)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.69)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.69)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.69)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.20)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.32)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.69)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.24)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.14)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.61)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.85)
57% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1322
\-------------------------

Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0712; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.52)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.04)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.49)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.95)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.95)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.95)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.95)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.95)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.84)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.82)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.86)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.89)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.89)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.89)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.89)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.52)
63% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1323
\-------------------------

Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0711; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.18)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.42)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.00)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.94)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.39)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.36)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.66)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.07)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1324
\-------------------------

Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0709; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.30)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.61)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.73)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.77)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.77)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.77)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.82)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.50)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.97)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.50)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 2.44)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.91)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.46)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1325
\-------------------------

Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0708; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.01)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.98)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.29)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.45)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.53)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.38)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.17)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.17)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.35)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.53)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded -0.13)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded -0.13)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded -0.13)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.40)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.32)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.80)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.98)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.78)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.55)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded -0.62)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1326
\-------------------------

Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0707; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.97)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.48)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.53)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.09)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.62)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.83)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.24)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.90)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.19)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.19)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.29)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.74)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.32)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.73)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 0.65)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.51)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.51)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.01)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.50)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1327
\-------------------------

Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0705; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.19)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.54)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.58)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.58)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.25)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.93)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove left instead of forward. (rewarded 1.09)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.83)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.93)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.93)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.93)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.93)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.93)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.33)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.00)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.00)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.00)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.00)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.11)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1328
\-------------------------

Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0704; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.45)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.05)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.51)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.17)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.85)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.61)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.05)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.05)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.05)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent drove right instead of forward. (rewarded 0.18)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.69)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent drove forward instead of left. (rewarded 1.04)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent drove forward instead of left. (rewarded 1.04)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent drove forward instead of left. (rewarded 1.04)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.24)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.49)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.49)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.49)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.49)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.09)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.09)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.05)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.35)
37% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1329
\-------------------------

Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0702; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.58)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.58)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.30)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.68)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.00)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.41)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.41)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.41)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.41)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.82)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.90)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.89)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.89)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.89)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.62)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.99)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.99)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.99)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.61)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 1.78)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 1.78)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.12)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.20)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.20)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.20)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.20)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.86)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.01)
51% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.01)
51% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.01)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.88)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.13)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.05)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.37)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.96)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.45)
34% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.45)
34% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.45)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.09)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.38)
29% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.38)
29% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.38)
29% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.38)
29% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.38)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.29)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.97)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.40)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.87)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent followed the waypoint right. (rewarded 1.77)
14% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent followed the waypoint right. (rewarded 1.77)
14% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent followed the waypoint right. (rewarded 1.77)
14% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent followed the waypoint right. (rewarded 1.77)
14% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent followed the waypoint right. (rewarded 1.77)
14% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent followed the waypoint right. (rewarded 1.77)
14% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent followed the waypoint right. (rewarded 1.77)
14% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent followed the waypoint right. (rewarded 1.77)
14% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent followed the waypoint right. (rewarded 1.77)
14% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent followed the waypoint right. (rewarded 1.77)
14% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent followed the waypoint right. (rewarded 1.77)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.34)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.59)
9% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1330
\-------------------------

Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0701; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.62)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.52)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.74)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.33)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.49)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.49)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.49)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.49)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.49)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.49)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.49)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.49)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.49)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.49)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.49)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.02)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.66)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.24)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.35)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.10)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.90)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1331
\-------------------------

Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0699; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.48)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.40)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.28)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.63)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1332
\-------------------------

Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0698; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.03)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.35)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.71)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.31)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.90)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.86)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.41)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.27)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.86)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.80)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.73)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.99)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1333
\-------------------------

Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0697; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent attempted driving left through a red light. (rewarded -9.11)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.67)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.66)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.66)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.66)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.66)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.66)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.16)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.94)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.94)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.17)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.01)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.01)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.01)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.74)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.74)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.73)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.06)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.06)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.31)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.33)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.20)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1334
\-------------------------

Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0695; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.80)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.33)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.20)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 2.44)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.02)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.02)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.17)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 1.05)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1335
\-------------------------

Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0694; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.34)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.95)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 0.42)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.75)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.73)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.27)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.40)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.19)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.19)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.19)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.62)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1336
\-------------------------

Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0693; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.92)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.65)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.61)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.61)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.61)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.61)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.61)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.61)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.61)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.61)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.61)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.20)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.84)
76% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1337
\-------------------------

Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0691; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.81)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.87)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.87)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.72)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.92)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.79)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.37)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.37)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.37)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.03)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.03)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.03)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.03)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.79)
63% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.79)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.95)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.95)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.95)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.95)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 2.43)
57% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1338
\-------------------------

Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0690; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.93)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.35)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.71)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.34)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.34)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.13)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.13)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.13)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.31)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.66)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.67)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.72)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.45)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.65)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.88)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.12)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.48)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1339
\-------------------------

Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0688; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.28)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.83)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.37)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.72)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.01)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.49)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.55)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.53)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.53)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.53)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.65)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.64)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.19)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.73)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.73)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.71)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.36)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 0.57)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1340
\-------------------------

Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0687; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.05)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 2.20)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.98)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.98)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.98)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.05)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.30)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.88)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.19)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.45)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.29)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.74)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.20)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.20)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.20)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.53)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.53)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.09)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.07)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.34)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.34)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.34)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.34)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.34)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.34)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.94)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.71)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.71)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.71)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.71)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.71)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.71)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.71)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.05)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1341
\-------------------------

Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0686; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.68)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.77)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.16)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.16)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.83)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.79)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.97)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.88)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1342
\-------------------------

Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0684; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.34)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.44)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.20)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.64)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.25)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.62)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.81)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.81)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.97)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.26)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.26)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.90)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.78)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.50)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.50)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.50)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.41)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.19)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 0.92)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.54)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.57)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.49)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.08)
33% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1343
\-------------------------

Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0683; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.86)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.34)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.34)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.83)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.09)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.09)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.09)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.76)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.98)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.52)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.43)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.67)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.71)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.60)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.33)
47% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1344
\-------------------------

Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0682; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent followed the waypoint left. (rewarded 1.31)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.94)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.94)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.94)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.94)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.90)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.00)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.36)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.36)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.64)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.22)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.75)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.30)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
66% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1345
\-------------------------

Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0680; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.27)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.54)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.55)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 2.85)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.03)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.03)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.37)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.66)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.47)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.33)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.33)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.33)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.33)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.33)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.33)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.33)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.04)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.16)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.16)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.14)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1346
\-------------------------

Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0679; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.81)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.80)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.80)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.80)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.80)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 1.73)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.55)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.52)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.23)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.72)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.41)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.44)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.44)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.73)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.92)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.62)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1347
\-------------------------

Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0677; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.90)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.47)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.65)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.13)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.47)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.26)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.26)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.26)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.12)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.26)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.22)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.76)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.19)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1348
\-------------------------

Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0676; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.42)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.41)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.24)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.29)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.49)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.49)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.49)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.49)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.85)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.24)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.33)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.74)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.69)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.50)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1349
\-------------------------

Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0675; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.73)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.75)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.93)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.93)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.93)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.93)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.93)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.93)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.93)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.93)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.14)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.23)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.88)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.86)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1350
\-------------------------

Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0673; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 1.58)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.77)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.47)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.29)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.97)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.37)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.37)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.37)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.37)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.37)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.60)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.36)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.36)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.62)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 1.25)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded -0.18)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 2.19)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 2.19)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 2.19)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 2.19)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove left instead of right. (rewarded 1.22)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.38)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.38)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.38)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.72)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.39)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.66)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1351
\-------------------------

Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0672; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.29)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.73)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.79)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.14)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1352
\-------------------------

Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0671; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.24)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.21)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.75)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.08)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 2.80)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 2.80)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.68)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.56)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.16)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.25)
67% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1353
\-------------------------

Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0669; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.41)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.57)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.60)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.74)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.85)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.85)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 2.82)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 2.82)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 2.82)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.20)
74% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1354
\-------------------------

Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0668; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 3.00)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.34)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.97)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.97)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.97)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.97)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.66)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.17)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.49)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.49)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.49)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.80)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.51)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.41)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.41)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.54)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.54)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.54)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.54)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.59)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 0.51)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.49)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.46)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.86)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.03)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.03)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.35)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.35)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.35)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.51)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.51)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.60)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1355
\-------------------------

Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0667; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.33)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.84)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.41)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.63)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.63)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.63)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.63)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.63)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.63)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.42)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.55)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.22)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.01)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.51)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.51)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.51)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 0.97)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1356
\-------------------------

Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0665; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.75)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.86)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.89)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.89)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.89)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.89)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.89)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.89)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.89)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.97)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.74)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 1.43)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 2.15)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.04)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.30)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.30)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.30)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.94)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove forward instead of left. (rewarded 0.78)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.97)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.25)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.25)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.25)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.25)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.10)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.05)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.05)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.05)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 0.41)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 0.41)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.02)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1357
\-------------------------

Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0664; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.41)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove right instead of forward. (rewarded 0.43)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.41)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.02)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.82)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.65)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.65)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.65)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.65)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.05)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.05)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.01)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.01)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.01)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.01)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.93)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.54)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.92)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.25)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.10)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1358
\-------------------------

Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0663; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.91)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.31)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.26)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.52)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.52)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.24)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.24)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.24)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.51)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.34)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.90)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.90)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.90)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.35)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 2.31)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 2.31)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.82)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.42)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 2.55)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 2.55)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 2.55)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 2.55)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 2.55)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.77)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.40)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.40)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.40)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.40)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.37)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.78)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1359
\-------------------------

Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0661; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 0.41)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.98)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.20)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.63)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.27)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.08)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.35)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.08)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.08)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.08)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.11)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.03)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1360
\-------------------------

Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0660; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.48)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.76)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.32)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.94)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.98)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.98)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.98)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.98)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.92)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.67)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.23)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.23)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.23)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.23)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.69)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.18)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.58)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.90)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.79)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1361
\-------------------------

Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0659; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.19)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.40)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.53)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.02)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.24)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.24)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.24)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.24)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.24)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.46)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.59)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.78)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.78)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.66)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.07)
67% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.07)
67% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.07)
67% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.07)
67% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.07)
67% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.07)
67% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.07)
67% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.07)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.15)
63% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1362
\-------------------------

Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0657; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.27)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.19)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.56)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.02)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.69)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.78)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.52)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.41)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.41)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.97)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.97)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.67)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.80)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.18)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.18)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.18)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.18)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.18)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.58)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.34)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 0.61)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1363
\-------------------------

Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0656; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.07)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.92)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'left')
Agent drove forward instead of left. (rewarded 0.63)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.74)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.17)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.86)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.33)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.28)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.72)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.87)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.54)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1364
\-------------------------

Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0655; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.32)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.28)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.60)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.95)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.32)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.17)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 2.80)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.72)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.29)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1365
\-------------------------

Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0653; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.74)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.30)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.57)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.49)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.49)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.49)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.49)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.49)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.49)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.49)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.49)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.49)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.49)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.49)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.33)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.63)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.32)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.41)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.63)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.59)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.95)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.61)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.44)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.44)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.44)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.44)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.09)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.61)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.61)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.61)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.61)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.88)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.27)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.75)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.29)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.29)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.29)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.52)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.73)
4% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1366
\-------------------------

Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0652; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.34)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.21)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.95)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.46)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.46)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.32)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.21)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.21)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.68)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.32)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1367
\-------------------------

Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0651; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.74)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.32)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.78)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.76)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.76)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.76)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.76)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.15)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.76)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.21)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.21)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.41)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.59)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.59)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.59)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.59)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.59)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.00)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.02)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.60)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.03)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.52)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1368
\-------------------------

Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0650; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.05)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.88)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.62)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded 0.00)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.07)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.07)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.07)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.07)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.07)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.61)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.68)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1369
\-------------------------

Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0648; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.58)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.53)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.67)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.46)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.70)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.77)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.77)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.77)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.75)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.07)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.92)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.92)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.92)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.87)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.87)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.29)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.29)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.07)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.93)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.22)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1370
\-------------------------

Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0647; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.18)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.78)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.78)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.76)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.82)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.17)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.17)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.17)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.45)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.06)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.46)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.75)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.76)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.19)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.19)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.19)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.19)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.19)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.24)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.77)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.77)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.77)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.82)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 2.63)
37% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1371
\-------------------------

Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0646; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.50)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.86)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.54)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.54)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.21)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.21)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.34)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.27)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.27)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 1.48)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.20)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.80)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.08)
67% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1372
\-------------------------

Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0644; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.89)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.69)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.96)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.72)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.15)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.05)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.05)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.87)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.87)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.87)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.87)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.76)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.41)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.41)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.88)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.88)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.88)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.25)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 1.42)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.79)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.24)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.86)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1373
\-------------------------

Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0643; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.40)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.16)
92% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.16)
92% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.16)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.70)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.31)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.62)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.12)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.70)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'right')
Agent drove left instead of forward. (rewarded 0.58)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.77)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1374
\-------------------------

Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0642; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.93)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.42)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.65)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.39)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.92)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.92)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.71)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.11)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.52)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.55)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.55)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.55)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.78)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.65)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.16)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1375
\-------------------------

Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0641; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.62)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.85)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.07)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.16)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.51)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.14)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.50)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.83)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.83)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.20)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.20)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.20)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.20)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 2.67)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1376
\-------------------------

Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0639; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.78)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.61)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.09)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.32)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.17)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.56)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.17)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.73)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.86)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1377
\-------------------------

Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0638; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 0.37)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.13)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.39)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.97)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.14)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.73)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.67)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1378
\-------------------------

Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0637; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.87)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.83)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.91)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.08)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.60)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.27)
73% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1379
\-------------------------

Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0635; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.26)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.75)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.51)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.51)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.51)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.51)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.26)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.36)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.81)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.81)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.81)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.24)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.38)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.38)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.38)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.65)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.65)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.65)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.65)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.00)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1380
\-------------------------

Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0634; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.45)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.67)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.67)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.67)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.67)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.11)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.11)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.11)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.11)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.16)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.09)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.18)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.37)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1381
\-------------------------

Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0633; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.82)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.57)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.31)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.71)
86% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.71)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.23)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.84)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.84)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.84)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.02)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.41)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.76)
66% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1382
\-------------------------

Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0632; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.00)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.83)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.19)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.89)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.26)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.26)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.26)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.88)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.09)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.52)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.52)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.52)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.52)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.40)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.61)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.54)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.14)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.39)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.67)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.60)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.02)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.48)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.56)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.56)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.56)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded -0.02)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.16)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.16)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.16)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.09)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.39)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.90)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.90)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.90)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.90)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.90)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.12)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1383
\-------------------------

Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0630; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.45)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.96)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.99)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.30)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.27)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.88)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.07)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.09)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.18)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.73)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.34)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.83)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.06)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.83)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.28)
43% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1384
\-------------------------

Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0629; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.19)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.25)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.83)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.12)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.21)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.21)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.21)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.21)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.02)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.27)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.98)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.98)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.98)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.57)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.57)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.57)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.82)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.10)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.55)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.96)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.89)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.51)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1385
\-------------------------

Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0628; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'right')
Agent followed the waypoint right. (rewarded 1.27)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.92)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.83)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.83)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.94)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.93)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.93)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.93)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.93)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.93)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.93)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.83)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.89)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.61)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.87)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.43)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.44)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.07)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.07)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.07)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.07)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.07)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.07)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', None)
Agent drove forward instead of right. (rewarded 0.20)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.51)
54% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.51)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.06)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.81)
49% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.81)
49% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.81)
49% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.81)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.74)
46% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.12)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.12)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.12)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.12)
43% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.12)
40% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.41)
37% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.43)
34% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.43)
34% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.43)
34% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.43)
34% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.43)
34% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.25)
31% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
29% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.25)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.25)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.25)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.25)
26% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.88)
23% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
20% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.60)
17% of time remaining to reach destination.

/-------------------
| Step 29 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.34)
14% of time remaining to reach destination.

/-------------------
| Step 30 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.67)
11% of time remaining to reach destination.

/-------------------
| Step 31 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.08)
9% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1386
\-------------------------

Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0627; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.30)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.55)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.70)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.70)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.70)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.70)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.70)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.70)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.70)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.58)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.22)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.61)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.61)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.78)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.40)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.59)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.15)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.94)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.35)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.45)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1387
\-------------------------

Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0625; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.31)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.18)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', 'right')
Agent properly idled at a red light. (rewarded 2.38)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.52)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.52)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.52)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.52)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.36)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.75)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.75)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.75)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.85)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.34)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.68)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.47)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.82)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.84)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 0.52)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.83)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.19)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1388
\-------------------------

Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0624; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.12)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.67)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.56)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 2.53)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 2.53)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 2.53)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.91)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.87)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.28)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.07)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.25)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.17)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.72)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.94)
32% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1389
\-------------------------

Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0623; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.29)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.25)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.48)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.83)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.90)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.74)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.25)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.90)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.90)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.90)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.22)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.72)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.29)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.28)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.03)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.61)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.61)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.61)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.61)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.14)
43% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1390
\-------------------------

Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0622; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.84)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.10)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.09)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.09)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.09)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.71)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.31)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.72)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.11)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1391
\-------------------------

Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0620; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.00)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.80)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.50)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.50)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.50)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.50)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.50)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.50)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.50)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.49)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.97)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.61)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.74)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.97)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.29)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.16)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1392
\-------------------------

Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0619; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.84)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.20)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.66)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 1.19)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 1.19)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.09)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.10)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.73)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.86)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.75)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.75)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.75)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.75)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.75)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.30)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.47)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.47)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.55)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.44)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.23)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded 0.05)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.61)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1393
\-------------------------

Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0618; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.88)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.05)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.64)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.79)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.01)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.58)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.35)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.60)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.81)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.28)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.43)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.43)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.43)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.41)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.97)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.72)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.32)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.09)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.09)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.09)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.28)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.61)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.28)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.28)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.20)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.96)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.15)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.21)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.31)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.96)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.96)
7% of time remaining to reach destination.

/-------------------
| Step 28 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 0.62)
3% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1394
\-------------------------

Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0617; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.94)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.97)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.35)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.41)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.60)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.60)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.94)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.27)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.61)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.95)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.39)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.43)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.11)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.49)
47% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1395
\-------------------------

Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0615; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.69)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.27)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.76)
91% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.76)
91% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.76)
91% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.76)
91% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.76)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.52)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.35)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.67)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.39)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.93)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.93)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.74)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.74)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.74)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.74)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.90)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.23)
54% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1396
\-------------------------

Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0614; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.64)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.21)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.28)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.22)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.95)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.97)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.63)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.63)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.63)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.28)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.19)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1397
\-------------------------

Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0613; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.71)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.57)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.44)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.44)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.99)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.99)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.99)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.99)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.96)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 1.35)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.30)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.80)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1398
\-------------------------

Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0612; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.05)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.15)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.34)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.34)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.76)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.76)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.76)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.27)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.14)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.52)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.18)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.88)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.94)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.17)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 2.10)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.53)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.15)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.06)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.56)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.01)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.01)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.65)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1399
\-------------------------

Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0611; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.39)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.46)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.84)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 1.03)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.58)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.69)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.91)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.56)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.83)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.14)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1400
\-------------------------

Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0609; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.02)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.18)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.55)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.55)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.55)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.24)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.24)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.73)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.49)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.62)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.62)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.31)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.13)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.76)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.08)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.83)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.83)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.83)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.83)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.83)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.00)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.65)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.65)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.22)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1401
\-------------------------

Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0608; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.65)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.32)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.94)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.27)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.39)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.39)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.39)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.39)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.39)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.84)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1402
\-------------------------

Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0607; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.46)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.13)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.00)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.24)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.09)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.79)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.79)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.48)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.87)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.15)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.21)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1403
\-------------------------

Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0606; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.86)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.10)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.55)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.14)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.48)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.96)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.35)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.28)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.65)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.67)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1404
\-------------------------

Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0604; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.11)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.38)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.38)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.38)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.38)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.38)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.34)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.06)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.05)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.71)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.50)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.39)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.39)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.39)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.87)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.72)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 0.97)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1405
\-------------------------

Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0603; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.16)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.15)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.63)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.23)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1406
\-------------------------

Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0602; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.64)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.34)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.14)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.63)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.29)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.29)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.49)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.49)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.66)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.55)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.16)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.47)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.47)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.82)
48% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1407
\-------------------------

Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0601; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.61)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.79)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.40)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.40)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.40)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.40)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.40)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.72)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.53)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.73)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.73)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.73)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.73)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.90)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.76)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1408
\-------------------------

Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0600; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.71)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.76)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.66)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.07)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.07)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.07)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.74)
75% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1409
\-------------------------

Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0598; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.70)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 1.87)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.40)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.43)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.88)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.68)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.68)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.16)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.02)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.18)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.46)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.46)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.46)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.76)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1410
\-------------------------

Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0597; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.91)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.27)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.69)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.52)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.52)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.52)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.52)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.64)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.29)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.68)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.68)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded -0.09)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded -0.09)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.24)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.38)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.38)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.38)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.38)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 1.59)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.43)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.22)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.22)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.22)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.34)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 0.71)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1411
\-------------------------

Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0596; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.93)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.64)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.23)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.88)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.48)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.00)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.00)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.00)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.00)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.03)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.55)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.55)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.98)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.98)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.98)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.10)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.12)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.70)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.97)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.39)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.39)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.77)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.64)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.64)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.64)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.64)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.42)
37% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1412
\-------------------------

Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0595; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 1.94)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 2.20)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.62)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.30)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.30)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', None)
Agent drove forward instead of right. (rewarded 0.48)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.71)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.75)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 2.30)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.34)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded -0.40)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.97)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.41)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.99)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.50)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.50)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.50)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.50)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.50)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1413
\-------------------------

Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0594; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.72)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.29)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.84)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.84)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.84)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.84)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.84)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.61)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.36)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.75)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.20)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.20)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.20)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.91)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
74% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
74% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
74% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.98)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.11)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.98)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.28)
63% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.10)
60% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.56)
57% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.92)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.92)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.92)
54% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.46)
51% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.70)
49% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.46)
46% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1414
\-------------------------

Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0592; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.44)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.53)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.77)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.87)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 0.94)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.95)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.95)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.95)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.90)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.52)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1415
\-------------------------

Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0591; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'right', 'right')
Agent followed the waypoint forward. (rewarded 2.86)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.87)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.58)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.93)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.93)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.93)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.93)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.46)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.70)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.03)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.16)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.13)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.15)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.15)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 1.30)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.09)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.90)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.98)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.98)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.98)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.48)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.10)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.10)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.10)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'right', 'left')
Agent drove right instead of left. (rewarded -0.22)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'right')
Agent followed the waypoint right. (rewarded 1.09)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1416
\-------------------------

Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0590; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'right')
Agent drove right instead of left. (rewarded 0.07)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.36)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.98)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.98)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.98)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.98)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.98)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.98)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.18)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.28)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.28)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.28)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.00)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.51)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1417
\-------------------------

Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0589; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 1.98)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.16)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.10)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.33)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.13)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.13)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.13)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.13)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.13)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.13)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.58)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.09)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 1.25)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.07)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.08)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.28)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.25)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.25)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.51)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.14)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.41)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.38)
5% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1418
\-------------------------

Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0588; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.61)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.50)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.33)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.40)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.89)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.94)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.94)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.94)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.14)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1419
\-------------------------

Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0587; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.20)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'right', 'right')
Agent properly idled at a red light. (rewarded 2.75)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.45)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.45)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.45)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.45)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.45)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.45)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.82)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.27)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.27)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.27)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.27)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.27)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.79)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.11)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.61)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.10)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1420
\-------------------------

Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0585; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.44)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.75)
94% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.19)
91% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.13)
89% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.34)
86% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.88)
83% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.87)
77% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.13)
74% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
71% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.34)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.34)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.34)
69% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.89)
66% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.55)
63% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1421
\-------------------------

Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0584; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 2.98)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.52)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.08)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.85)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.17)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.17)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.12)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.95)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.95)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.95)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.95)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.95)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.83)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.04)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.04)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.87)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.87)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.74)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.19)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.22)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.09)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1422
\-------------------------

Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0583; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.10)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.33)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.35)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.76)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.58)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.13)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.18)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.81)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.81)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.37)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.95)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.26)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.75)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1423
\-------------------------

Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0582; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.12)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.49)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.70)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.03)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.03)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.03)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.03)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.80)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.80)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.80)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.94)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.91)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.91)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.91)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.91)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.91)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 2.65)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.65)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.00)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.00)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.01)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.02)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.29)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.40)
37% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1424
\-------------------------

Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0581; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.11)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.79)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.79)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.79)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.79)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.79)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.47)
84% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.47)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.35)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.89)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.70)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.69)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.53)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.53)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.53)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.53)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.65)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.84)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.08)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.08)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.08)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.08)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.06)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.34)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.34)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.70)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.36)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.36)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 0.62)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.50)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.50)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.50)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.50)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.50)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.09)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1425
\-------------------------

Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0580; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.84)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'right')
Agent followed the waypoint right. (rewarded 1.00)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.65)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.77)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.77)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.77)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.61)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.45)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.73)
72% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1426
\-------------------------

Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0578; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 1.94)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.43)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.27)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.27)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.27)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.27)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.27)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.27)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.13)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.26)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.62)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.85)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.90)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.90)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.90)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.90)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.90)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.68)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1427
\-------------------------

Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0577; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.05)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.54)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.16)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.16)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.16)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.16)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.16)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.16)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.16)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.22)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.59)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.70)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.40)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.21)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.15)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.15)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 1.00)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.77)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.54)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.54)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.81)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.81)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.81)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.81)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.81)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.81)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.81)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.09)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.54)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.78)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1428
\-------------------------

Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0576; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.11)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.98)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.47)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.97)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1429
\-------------------------

Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0575; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.93)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.09)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.61)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.61)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.61)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.85)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.16)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.04)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1430
\-------------------------

Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0574; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 0.32)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.87)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.21)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.02)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.10)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.04)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.04)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.04)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.04)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.04)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.04)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.04)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.04)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.94)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', 'left')
Agent followed the waypoint left. (rewarded 1.03)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1431
\-------------------------

Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0573; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.78)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.66)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 1.73)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.72)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.55)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.23)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.11)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 2.08)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.40)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.93)
57% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1432
\-------------------------

Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0572; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.96)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.21)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.57)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.57)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.57)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.57)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.39)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.60)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.60)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.60)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.98)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.18)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.60)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.60)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.60)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.92)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.88)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1433
\-------------------------

Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0570; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.10)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.53)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.46)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.46)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.46)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.46)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.64)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.27)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.27)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.27)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.27)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.20)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.78)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.78)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.78)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.78)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.38)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.37)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.62)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.62)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.62)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.62)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.62)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 1.66)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.81)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.57)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.07)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.91)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.09)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.09)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.79)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.45)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1434
\-------------------------

Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0569; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 1.26)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.65)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.35)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.79)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.51)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.58)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.61)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.77)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.77)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.85)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.01)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1435
\-------------------------

Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0568; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.63)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.74)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.67)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.64)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.64)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.40)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.43)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.27)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.27)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.36)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.48)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.48)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.48)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.49)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.09)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.77)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.63)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.63)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.63)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.63)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.63)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.38)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.25)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.58)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.49)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.88)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.05)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.28)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
10% of time remaining to reach destination.

/-------------------
| Step 27 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.05)
7% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1436
\-------------------------

Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0567; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.01)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.46)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.34)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.34)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.49)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 1.81)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.26)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.86)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.07)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.07)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.07)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.57)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.60)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.74)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.00)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.00)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.64)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.95)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.30)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove left instead of forward. (rewarded -0.52)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove left instead of forward. (rewarded -0.52)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove left instead of forward. (rewarded -0.52)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove left instead of forward. (rewarded -0.52)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.43)
5% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.43)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.04)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1437
\-------------------------

Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0566; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 1.38)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.30)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.26)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.73)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.45)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.77)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.93)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 1.59)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.04)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.04)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.04)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.89)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.36)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1438
\-------------------------

Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0565; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.46)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.88)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.57)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.57)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.57)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.57)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.57)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.57)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.57)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.50)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.40)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.68)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.30)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.76)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.81)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.06)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.98)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1439
\-------------------------

Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0564; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.54)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.60)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.60)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.60)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.60)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.60)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.60)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.60)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.60)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.80)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.80)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.96)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.30)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.04)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.46)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.46)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.46)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.37)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.90)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.90)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.90)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.11)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.63)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.63)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.73)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1440
\-------------------------

Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0562; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.93)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.54)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.63)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent drove forward instead of left. (rewarded 0.08)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.67)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove forward instead of left. (rewarded 1.08)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.56)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1441
\-------------------------

Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0561; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.75)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.75)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.75)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.75)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.75)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.75)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.00)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.34)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.86)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.27)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.75)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.75)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.84)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.42)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.21)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1442
\-------------------------

Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0560; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.93)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.28)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.88)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.08)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.65)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.73)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.16)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.19)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.75)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.75)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.75)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.13)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1443
\-------------------------

Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0559; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.59)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.27)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.86)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.45)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.45)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.45)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.45)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.45)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.45)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.15)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1444
\-------------------------

Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0558; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.87)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.15)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.23)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.23)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.03)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.69)
83% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.69)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.75)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.75)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.75)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.75)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.28)
77% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1445
\-------------------------

Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0557; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.15)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.44)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.04)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.04)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.04)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.72)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.72)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.08)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.84)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.25)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.00)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded 1.67)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.74)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.90)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.22)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.01)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.91)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.91)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.91)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.91)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.35)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.65)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.65)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.65)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.65)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.65)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.19)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.35)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.67)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 1.77)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1446
\-------------------------

Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0556; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.53)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.54)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.20)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.32)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.93)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.50)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.50)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.50)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.50)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.55)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.44)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.27)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.38)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1447
\-------------------------

Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0555; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.10)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.74)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.79)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.11)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.50)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.50)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.15)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.91)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.96)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.96)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.96)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.96)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.44)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.38)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 0.69)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.07)
35% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1448
\-------------------------

Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0554; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.41)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.14)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.95)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.47)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.39)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.72)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.18)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.18)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.18)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.98)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.95)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.95)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.95)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.95)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.49)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1449
\-------------------------

Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0552; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.22)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.50)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.69)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
70% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1450
\-------------------------

Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0551; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.55)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.10)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.03)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.15)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.15)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.15)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.15)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.15)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.15)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.15)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.56)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.89)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.89)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.89)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.89)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.89)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.55)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.41)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.41)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.41)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.41)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.16)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.06)
52% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1451
\-------------------------

Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0550; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.54)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.76)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.69)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.69)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.69)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.67)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.37)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.37)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.77)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.95)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.11)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.11)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.74)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1452
\-------------------------

Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0549; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.89)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.89)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.66)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.41)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.41)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.03)
76% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1453
\-------------------------

Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0548; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.33)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.86)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.54)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.54)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.54)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.54)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.54)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.54)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.30)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.76)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.93)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1454
\-------------------------

Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0547; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.90)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.87)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.17)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.99)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.96)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.03)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.03)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.39)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 1.20)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.15)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.93)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1455
\-------------------------

Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0546; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.88)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.50)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.31)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.22)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1456
\-------------------------

Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0545; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.19)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.51)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.69)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.20)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.51)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.66)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.58)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.58)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.41)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.16)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.62)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.37)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1457
\-------------------------

Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0544; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.99)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.95)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1458
\-------------------------

Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0543; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.71)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.89)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.08)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.72)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.85)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.85)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent drove right instead of forward. (rewarded 1.89)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.57)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.57)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.57)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.57)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.39)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.66)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.08)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.54)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.54)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.54)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.54)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.25)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.32)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.09)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'right', None)
Agent drove right instead of left. (rewarded 1.30)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.03)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.97)
27% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1459
\-------------------------

Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0541; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent drove right instead of forward. (rewarded 0.54)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.07)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.72)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.04)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.58)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.28)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1460
\-------------------------

Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0540; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.29)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.96)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.13)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.13)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.13)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.13)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.93)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.55)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.60)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.13)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.61)
67% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1461
\-------------------------

Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0539; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.47)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.55)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.92)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.03)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.84)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.11)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.71)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.07)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.96)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.51)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.74)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.17)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.27)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.27)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.27)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded -0.28)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded -0.28)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.34)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.34)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.34)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.34)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.36)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.70)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.66)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1462
\-------------------------

Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0538; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'right')
Agent drove right instead of forward. (rewarded 1.60)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.24)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.31)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.31)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.67)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.44)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.44)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.44)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.23)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.23)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.39)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.40)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.25)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.04)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.04)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.59)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1463
\-------------------------

Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0537; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.09)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.30)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.29)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.25)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.09)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.48)
72% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1464
\-------------------------

Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0536; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.42)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.96)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.86)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.48)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.48)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.48)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.48)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.97)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.06)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.69)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.83)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent drove right instead of left. (rewarded 1.34)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.15)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 2.20)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 2.20)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 2.20)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 2.20)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.91)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.29)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.19)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.71)
10% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.03)
5% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.87)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Training trial 1465
\-------------------------

Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0535; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.57)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.91)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.39)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.93)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.08)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.88)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.20)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.79)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.79)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.79)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.79)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.79)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.80)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1466
\-------------------------

Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0534; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.34)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.00)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.33)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.40)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.76)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.76)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.02)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.35)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.37)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.37)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.37)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.07)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.31)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.31)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.31)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.31)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.43)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 1.49)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.08)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.44)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.48)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.31)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.58)
20% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1467
\-------------------------

Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0533; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.49)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.10)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.58)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.53)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.04)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.44)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.40)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.10)
68% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1468
\-------------------------

Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0532; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.58)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.70)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.57)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.38)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.38)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.38)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.76)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.24)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.24)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.24)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.69)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.58)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.58)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.58)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.73)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1469
\-------------------------

Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0531; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.42)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.66)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.18)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.95)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.33)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.00)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.78)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.00)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.70)
64% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1470
\-------------------------

Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0530; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.06)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.69)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.00)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.00)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.78)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.78)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.78)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.78)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.78)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.58)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.23)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.15)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1471
\-------------------------

Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0529; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.49)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.63)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.63)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.63)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.63)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.63)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.65)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.51)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.92)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.81)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.13)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.47)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.51)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.93)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.97)
36% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1472
\-------------------------

Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0528; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.69)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.75)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.70)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.19)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.83)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.04)
72% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1473
\-------------------------

Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0527; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.80)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.80)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.28)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.28)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.28)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.68)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.70)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.16)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 1.62)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.94)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.38)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.04)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1474
\-------------------------

Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0525; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.35)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.23)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.52)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.90)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.41)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.30)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'right', 'right')
Agent drove left instead of forward. (rewarded 0.20)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.00)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.00)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.43)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.36)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.35)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.86)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.24)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.24)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.24)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.98)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.92)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1475
\-------------------------

Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0524; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.43)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.56)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.87)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.52)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.52)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.52)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.52)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.19)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.55)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.24)
72% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1476
\-------------------------

Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0523; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.73)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.27)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.86)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.86)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.86)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.86)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.86)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.86)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.86)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.86)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.86)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.86)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.13)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.57)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.41)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.59)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.22)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.69)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.29)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.86)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.51)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.80)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.10)
53% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1477
\-------------------------

Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0522; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.66)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.57)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.55)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.12)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.12)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.12)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.12)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.12)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.12)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.12)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.12)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.97)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.79)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.87)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.51)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.93)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.17)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1478
\-------------------------

Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0521; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.63)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.01)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.62)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.06)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.10)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.71)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.71)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.71)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.15)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.15)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.78)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.21)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.21)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 1.18)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.24)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.73)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.62)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.44)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.80)
28% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1479
\-------------------------

Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0520; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', 'right')
Agent drove right instead of forward. (rewarded 1.84)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.88)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.52)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.52)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.52)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.52)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.52)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.12)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.12)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.12)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.00)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.61)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.68)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1480
\-------------------------

Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0519; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.55)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.77)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.09)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.54)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.29)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.29)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.98)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.19)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.49)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.49)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.49)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.20)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.51)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1481
\-------------------------

Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0518; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.26)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.99)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.62)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.62)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.62)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.62)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.62)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.62)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.62)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.62)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.62)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.62)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.62)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.78)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.11)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.11)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.11)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.11)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.11)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.11)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.11)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.82)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.28)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.12)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.65)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.78)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.33)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.85)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent drove right instead of forward. (rewarded 0.56)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.55)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.80)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.80)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.80)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.80)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.80)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.76)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.34)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.34)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.34)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.34)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.34)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.34)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.34)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.34)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.83)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.75)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.55)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.15)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.49)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.49)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.73)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.50)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1482
\-------------------------

Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0517; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.81)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.69)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.99)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.96)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.31)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.33)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.45)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.48)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.45)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.57)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.85)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.44)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.52)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.25)
30% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1483
\-------------------------

Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0516; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.58)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.13)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 2.20)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.96)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.96)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.96)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.63)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.28)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.57)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.91)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1484
\-------------------------

Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0515; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.16)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.41)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.94)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.81)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.65)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.88)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.16)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 2.14)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.58)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1485
\-------------------------

Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0514; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.32)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.64)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.41)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.21)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.48)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.84)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1486
\-------------------------

Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0513; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.08)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.31)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.63)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.72)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.66)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.02)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.33)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.44)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.16)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.16)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.16)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.25)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.75)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.19)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.27)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.76)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.70)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.70)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.44)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.44)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.44)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.44)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.44)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.06)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.38)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
27% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.44)
23% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.68)
20% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.23)
17% of time remaining to reach destination.

/-------------------
| Step 25 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.22)
13% of time remaining to reach destination.

/-------------------
| Step 26 Results
\-------------------

Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 0.50)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1487
\-------------------------

Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0512; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.67)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.59)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.85)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.84)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.89)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.89)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.76)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.76)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.76)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.35)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.00)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 2.19)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.81)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.49)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.10)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.62)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.49)
15% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1488
\-------------------------

Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0511; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.16)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.70)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.86)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.12)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.12)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.12)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.12)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.12)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.42)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.20)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.96)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.37)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.47)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.47)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.97)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.48)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.48)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.48)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.48)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.28)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.60)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.98)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.19)
36% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1489
\-------------------------

Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0510; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.10)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.91)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.11)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.08)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1490
\-------------------------

Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0509; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.92)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.31)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.09)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.35)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1491
\-------------------------

Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0508; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.59)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.10)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.95)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.95)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.37)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.22)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.08)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.85)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 0.91)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.70)
63% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.70)
63% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.70)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.22)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1492
\-------------------------

Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0507; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.21)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.27)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.74)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.70)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.89)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.46)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.58)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.32)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.09)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.95)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.95)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.65)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.79)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.79)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.79)
25% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.00)
20% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.43)
15% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.12)
10% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1493
\-------------------------

Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0506; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.72)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.58)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.37)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.79)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.67)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.92)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.61)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.67)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.67)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.67)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.67)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.67)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.17)
56% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1494
\-------------------------

Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0505; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.85)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.22)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.04)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.57)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.45)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent drove forward instead of left. (rewarded 0.48)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.09)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 0.51)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.06)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 1.47)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 1.47)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 1.47)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 1.47)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.30)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.35)
40% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.85)
35% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.44)
30% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.57)
25% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1495
\-------------------------

Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0504; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove forward instead of left. (rewarded 0.23)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.81)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.26)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.70)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.92)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.92)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.92)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.92)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.92)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.42)
45% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1496
\-------------------------

Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0503; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.22)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.57)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.61)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.16)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.14)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.95)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.95)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.88)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.02)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.36)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.45)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.45)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.05)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.39)
44% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1497
\-------------------------

Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0502; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.03)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.40)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.60)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.60)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.60)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.77)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.90)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.49)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 1.04)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.44)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.90)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.90)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.12)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.10)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1498
\-------------------------

Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0501; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.05)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.29)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.10)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.15)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.79)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.66)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.76)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.40)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.24)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.33)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.42)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent drove right instead of left. (rewarded 0.32)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent drove right instead of left. (rewarded 0.32)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent drove right instead of left. (rewarded 0.32)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent drove right instead of left. (rewarded 0.32)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.65)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.19)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('left', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 1.84)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.65)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.65)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.65)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.80)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.98)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.98)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.98)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.98)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.25)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.89)
12% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Training trial 1499
\-------------------------

Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020
Simulating trial. . . 
epsilon = 0.0500; alpha = 0.0020

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.45)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.25)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.05)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.34)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.54)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.54)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.54)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.54)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.62)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.83)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.83)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.83)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
50% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.82)
45% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.59)
40% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Testing trial 1
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.11)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.80)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.74)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.00)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.89)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.29)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.65)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.68)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.11)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.17)
60% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.46)
56% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.01)
52% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.72)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.72)
48% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.02)
44% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.98)
40% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.93)
36% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.23)
32% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.25)
28% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
24% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded -0.41)
20% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.29)
16% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.04)
12% of time remaining to reach destination.

/-------------------
| Step 22 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.82)
8% of time remaining to reach destination.

/-------------------
| Step 23 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.44)
4% of time remaining to reach destination.

/-------------------
| Step 24 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.80)
0% of time remaining to reach destination.

Trial Aborted!
Agent did not reach the destination.

/-------------------------
| Testing trial 2
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.32)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.31)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.02)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.46)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.86)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.01)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.01)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.01)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.63)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.69)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.42)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.18)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.10)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.01)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 1.24)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.51)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.30)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Testing trial 3
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent followed the waypoint left. (rewarded 1.03)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.75)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.40)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.60)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.60)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.60)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.60)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.60)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.17)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.82)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.26)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.07)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.18)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.18)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.18)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.18)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.06)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.96)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.83)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.01)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.90)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.90)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.90)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.42)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.48)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.16)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.16)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.44)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.44)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.72)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.63)
33% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Testing trial 4
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.33)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.99)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.37)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.57)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Testing trial 5
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.08)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.33)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.89)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.38)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.38)
80% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.00)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.00)
76% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.40)
72% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.02)
68% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.65)
64% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.06)
60% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Testing trial 6
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.87)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 2.71)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.84)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.84)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.06)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 2.12)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.67)
55% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Testing trial 7
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.51)
97% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.53)
93% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.09)
90% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.26)
87% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.78)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.78)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.78)
83% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.21)
80% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.01)
77% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.74)
73% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
70% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
67% of time remaining to reach destination.

/-------------------
| Step 10 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.09)
63% of time remaining to reach destination.

/-------------------
| Step 11 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.13)
60% of time remaining to reach destination.

/-------------------
| Step 12 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'left', 'right')
Agent drove left instead of right. (rewarded 0.57)
57% of time remaining to reach destination.

/-------------------
| Step 13 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.09)
53% of time remaining to reach destination.

/-------------------
| Step 14 Results
\-------------------

Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
50% of time remaining to reach destination.

/-------------------
| Step 15 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
47% of time remaining to reach destination.

/-------------------
| Step 16 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.10)
43% of time remaining to reach destination.

/-------------------
| Step 17 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.60)
40% of time remaining to reach destination.

/-------------------
| Step 18 Results
\-------------------

Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.30)
37% of time remaining to reach destination.

/-------------------
| Step 19 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.81)
33% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.81)
33% of time remaining to reach destination.

/-------------------
| Step 20 Results
\-------------------

Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.23)
30% of time remaining to reach destination.

/-------------------
| Step 21 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.69)
27% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Testing trial 8
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.41)
96% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.14)
92% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.01)
88% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.92)
84% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.95)
80% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Testing trial 9
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.78)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.62)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.58)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.24)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.24)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.24)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.24)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.24)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.24)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.72)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.72)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.72)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.15)
65% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

/-------------------------
| Testing trial 10
\-------------------------

Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . . 
epsilon = 0.0000; alpha = 0.0000

/-------------------
| Step 0 Results
\-------------------

Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.01)
95% of time remaining to reach destination.

/-------------------
| Step 1 Results
\-------------------

Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.94)
90% of time remaining to reach destination.

/-------------------
| Step 2 Results
\-------------------

Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.56)
85% of time remaining to reach destination.

/-------------------
| Step 3 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.16)
80% of time remaining to reach destination.

/-------------------
| Step 4 Results
\-------------------

Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.84)
75% of time remaining to reach destination.

/-------------------
| Step 5 Results
\-------------------

Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
70% of time remaining to reach destination.

/-------------------
| Step 6 Results
\-------------------

Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.84)
65% of time remaining to reach destination.

/-------------------
| Step 7 Results
\-------------------

Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.04)
60% of time remaining to reach destination.

/-------------------
| Step 8 Results
\-------------------

Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.49)
55% of time remaining to reach destination.

/-------------------
| Step 9 Results
\-------------------

Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.51)
50% of time remaining to reach destination.

Trial Completed!
Agent reached the destination.

Simulation ended. . . 
In [3]:
# Load the 'sim_improved-learning' file from the improved Q-Learning simulation
vs.plot_trials('sim_improved-learning.csv')
print "decay function is :ϵ=e−at, alpha=0.002,epsilon=0.5"
decay function is :ϵ=e−at, alpha=0.002,epsilon=0.5

Answer:

  • An epsilon = e^(-at) decay function was used, with epsilon starting at 1. This was chosen because it is important to have a great wealth of learning early on, with a gradual decline into exploration.
  • Approximately 1400 trials were needed before testing was completed.
  • An epsilon tolerance of 0.05 was used, and an alpha of 0.002. The very low alpha was used to allow the model sufficient trials to train with, and caused the 1400 of trials. A large number of trials is important to increase the safety rating to a maximum. As a result of the small alpha, a smaller tolerance is more appropriate - otherwise a large proportion of trials towards the end of the decay would be missed.
  • A large improvement has been made from the Q-Learner from the previous section.
  • The Safety Rating has increased from an F to an A+
  • The Reliability Rating has increased from an F to an A.
  • These are both significant improvements that would mean the Smartcab safe for deployment.
  • The very good rating in smartcab performance indicates that the Q-Learner has successfully learned the policy
  • In a real-world driving environment, traffic rules are complex and only 344 combination of states we have seen in this example. In real work there would be more no of states and combination also increase which results in increase in Q-table.

Define an Optimal Policy

Sometimes, the answer to the important question "what am I trying to get my agent to learn?" only has a theoretical answer and cannot be concretely described. Here, however, you can concretely define what it is the agent is trying to learn, and that is the U.S. right-of-way traffic laws. Since these laws are known information, you can further define, for each state the Smartcab is occupying, the optimal action for the driving agent based on these laws. In that case, we call the set of optimal state-action pairs an optimal policy. Hence, unlike some theoretical answers, it is clear whether the agent is acting "incorrectly" not only by the reward (penalty) it receives, but also by pure observation. If the agent drives through a red light, we both see it receive a negative reward but also know that it is not the correct behavior. This can be used to your advantage for verifying whether the policy your driving agent has learned is the correct one, or if it is a suboptimal policy.

Question 8

  1. Please summarize what the optimal policy is for the smartcab in the given environment. What would be the best set of instructions possible given what we know about the environment? You can explain with words or a table, but you should thoroughly discuss the optimal policy.

  2. Next, investigate the 'sim_improved-learning.txt' text file to see the results of your improved Q-Learning algorithm. For each state that has been recorded from the simulation, is the policy (the action with the highest value) correct for the given state? Are there any states where the policy is different than what would be expected from an optimal policy?

  3. Provide a few examples from your recorded Q-table which demonstrate that your smartcab learned the optimal policy. Explain why these entries demonstrate the optimal policy.

  4. Try to find at least one entry where the smartcab did not learn the optimal policy. Discuss why your cab may have not learned the correct policy for the given state.

Be sure to document your state dictionary below, it should be easy for the reader to understand what each state represents.

Answer:

  • The state space defined in agent.py has following state
  • ( waypoint,light,oncoming,right,left)
  • the optimal policy for the algorithm should have a following important features:
  • Any forward or left turn action in a red light state should have the largest negative penalty. This would mean the agent has learnt the most basic traffic rule.
  • eg. the agent in state (right, 'red' , forward) and (right, 'red', 'left') should not have much difference and will likely just stay idle.
  • most of the states in the 'green' light should encourage movement, but certain conditions should be satisfied.
  • if the oncoming traffic is in a direction which will cause a collision, the action should be avoided.
  • if the oncoming traffic is in the same direction as the waypoint, the waypoint direction should be given preference.
  • eg. the agent in a state (forward, green, forward) should be ok with going forward, since it will not cause a collision and will get them closer to the destination
  • the agent should however learn the left turn rule, where it is not allowed to turn left if the oncoming traffic is turning right.
  • Optimal policy
  • eg. the agent in a state (left, green, right) should have a heavy penalty on the left action. ('forward', 'green', 'forward', 'forward', 'forward') -- forward : 0.02 -- right : 0.00 -- None : -0.01 -- left : -0.04 This matches reason stated above that agent in the state of green light should encourage movement and should take left only if there is no oncoming traffic, in this policy the agent rewarded high for forward action and negative rewards for staying idle and taking left if oncoming traffic in green signal. ('forward', 'red', None, None, None) -- forward : -2.28 -- right : 0.14 -- None : 1.66 -- left : -2.11 This policy shows that agent learnt the US traffic rule because it has given heavy penality for forward action in red light and taking left in red . Given positive reward for right action in red signal when there is no traffic in left. Again staying idle in red light when way point is forward is given high reward. It clearly shows that the agent has learnt well.

Optional: Future Rewards - Discount Factor, 'gamma'

Curiously, as part of the Q-Learning algorithm, you were asked to not use the discount factor, 'gamma' in the implementation. Including future rewards in the algorithm is used to aid in propagating positive rewards backwards from a future state to the current state. Essentially, if the driving agent is given the option to make several actions to arrive at different states, including future rewards will bias the agent towards states that could provide even more rewards. An example of this would be the driving agent moving towards a goal: With all actions and rewards equal, moving towards the goal would theoretically yield better rewards if there is an additional reward for reaching the goal. However, even though in this project, the driving agent is trying to reach a destination in the allotted time, including future rewards will not benefit the agent. In fact, if the agent were given many trials to learn, it could negatively affect Q-values!

Optional Question 9

There are two characteristics about the project that invalidate the use of future rewards in the Q-Learning algorithm. One characteristic has to do with the Smartcab itself, and the other has to do with the environment. Can you figure out what they are and why future rewards won't work for this project?

Answer:

Note: Once you have completed all of the code implementations and successfully answered each question above, you may finalize your work by exporting the iPython Notebook as an HTML document. You can do this by using the menu above and navigating to
File -> Download as -> HTML (.html). Include the finished document along with this notebook as your submission.